W0909 15:07:04.085000 139786888062784 torch/fx/experimental/symbolic_shapes.py:4449] [0/0] xindex is not in var_ranges, defaulting to unknown range. V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2169] [0/0] [__guards] GUARDS: V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] TREE_GUARD_MANAGER: V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] +- RootGuardManager V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- DEFAULT_DEVICE: utils_device.CURRENT_DEVICE == None # _dynamo/output_graph.py:460 in init_ambient_guards V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GLOBAL_STATE: ___check_global_state() V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['self'], accessed_by=DictGetItemGuardAccessor(self) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- ID_MATCH: ___check_obj_id(L['self'], 139777664825136) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=L['self'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=L['self'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(L['self'].training, 139786895605728) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=L['self']._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].norm_out, accessed_by=DictGetItemGuardAccessor(norm_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out, 139777661036496) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].norm_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].norm_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.training, 139786895605728) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm, 139777661037408) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu, 139777661037312) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear, 139777661037360) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out, 139777661037264) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out.training, 139786895605728) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].pos_embed, accessed_by=DictGetItemGuardAccessor(pos_embed) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed, 139777664824800) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].pos_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].pos_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed.training, 139786895605728) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim, accessed_by=DictGetItemGuardAccessor(axes_dim) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].pos_embed.axes_dim, 139786895700800) # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- LENGTH_CHECK: len(L['self'].pos_embed.axes_dim) == 3 # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[0], accessed_by=TupleGetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[0] == 16 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[1], accessed_by=TupleGetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[1] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[2], accessed_by=TupleGetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[2] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].x_embedder, accessed_by=DictGetItemGuardAccessor(x_embedder) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder, 139777664824032) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].x_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].x_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder.training, 139786895605728) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].time_text_embed, accessed_by=DictGetItemGuardAccessor(time_text_embed) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed, 139777664824848) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].time_text_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.training, 139786895605728) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj, accessed_by=DictGetItemGuardAccessor(time_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj, 139777664824752) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.training, 139786895605728) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.scale, accessed_by=DictGetItemGuardAccessor(scale) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.scale == 1 # scale=self.scale, # diffusers/src/diffusers/models/embeddings.py:769 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.num_channels, accessed_by=DictGetItemGuardAccessor(num_channels) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.num_channels == 256 # self.num_channels, # diffusers/src/diffusers/models/embeddings.py:766 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.flip_sin_to_cos, accessed_by=DictGetItemGuardAccessor(flip_sin_to_cos) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.flip_sin_to_cos, 139786895605696) # flip_sin_to_cos=self.flip_sin_to_cos, # diffusers/src/diffusers/models/embeddings.py:767 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.downscale_freq_shift, accessed_by=DictGetItemGuardAccessor(downscale_freq_shift) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.downscale_freq_shift == 0 # downscale_freq_shift=self.downscale_freq_shift, # diffusers/src/diffusers/models/embeddings.py:768 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder, accessed_by=DictGetItemGuardAccessor(text_embedder) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder, 139777664824416) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.text_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.training, 139786895605728) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1, accessed_by=DictGetItemGuardAccessor(act_1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1, 139777664824128) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1.training, 139786895605728) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1, 139777664824176) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1.training, 139786895605728) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2, 139777664824080) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2.training, 139786895605728) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder, accessed_by=DictGetItemGuardAccessor(guidance_embedder) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder, 139777664824560) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.guidance_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.training, 139786895605728) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1, 139777664824512) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2, 139777664824464) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder, accessed_by=DictGetItemGuardAccessor(timestep_embedder) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder, 139777664824704) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.timestep_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.training, 139786895605728) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1, 139777664824656) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2, 139777664824608) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].context_embedder, accessed_by=DictGetItemGuardAccessor(context_embedder) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder, 139777664824320) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].context_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].context_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder.training, 139786895605728) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].transformer_blocks, accessed_by=DictGetItemGuardAccessor(transformer_blocks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks, 139777664823936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0], 139777664823984) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff, 139777664822544) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net, 139777664822304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0], 139777664822352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139777664822256) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1], 139777664822208) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2], 139777664822160) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn, 139777664823360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k, 139777664823168) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q, 139777664823072) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v, 139777664822976) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k, 139777664823120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.weight, 139777659850240) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q, 139777664823216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.weight, 139777660412416) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out, 139777664822784) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0], 139777664822736) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1], 139777664822688) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj, 139777664822928) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj, 139777664822832) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj, 139777664822880) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out, 139777664822640) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k, 139777664822496) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.weight, 139777664182976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q, 139777664822592) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.weight, 139777664182016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.processor, 139777664823408) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1, 139777664823840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm, 139777664823648) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu, 139777664823744) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear, 139777664823696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2, 139777664822448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context, 139777664822112) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net, 139777664821968) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0], 139777664822016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj, 139777664821920) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1], 139777664821824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2], 139777664821776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context, 139777664823600) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm, 139777664823456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu, 139777664823552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear, 139777664823504) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context, 139777664822400) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1], 139777664823888) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff, 139777664820480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net, 139777664820240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0], 139777664820288) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj, 139777664820192) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1], 139777664820144) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2], 139777664820096) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn, 139777664821248) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k, 139777664821104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q, 139777664821008) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v, 139777664820912) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k, 139777664821056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.weight, 139777660113664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q, 139777664821152) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.weight, 139777659965328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out, 139777664820720) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0], 139777664820672) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1], 139777664820624) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj, 139777664820864) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj, 139777664820768) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj, 139777664820816) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out, 139777664820576) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k, 139777664820432) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.weight, 139777659965408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q, 139777664820528) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.weight, 139777659965488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.processor, 139777664821296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1, 139777664821728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm, 139777664821584) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu, 139777664821680) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear, 139777664821632) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2, 139777664820384) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context, 139777664820048) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net, 139777664819904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0], 139777664819952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj, 139777664819856) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1], 139777664819760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2], 139777664819712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context, 139777664821536) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm, 139777664821344) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu, 139777664821440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear, 139777664821392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context, 139777664820336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2], 139777664822064) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff, 139777664818416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net, 139777664818176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0], 139777664818224) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj, 139777664818128) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1], 139777664818080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2], 139777664818032) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn, 139777664819184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k, 139777664819040) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q, 139777664818944) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v, 139777664818848) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k, 139777664818992) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.weight, 139777664567632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q, 139777664819088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.weight, 139777664569792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out, 139777664818656) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0], 139777664818608) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1], 139777664818560) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj, 139777664818800) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj, 139777664818704) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj, 139777664818752) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out, 139777664818512) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k, 139777664818368) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.weight, 139777664568672) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q, 139777664818464) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.weight, 139777664567552) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.processor, 139777664819232) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1, 139777664819664) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm, 139777664819520) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu, 139777664819616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear, 139777664819568) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2, 139777664818320) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context, 139777664817984) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net, 139777664817840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0], 139777664817888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj, 139777664817792) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1], 139777664817696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2], 139777664817648) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context, 139777664819472) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm, 139777664819280) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu, 139777664819376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear, 139777664819328) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context, 139777664818272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3], 139777664820000) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff, 139777664816352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net, 139777664816112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0], 139777664816160) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj, 139777664816064) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1], 139777664816016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2], 139777664815968) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn, 139777664817120) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k, 139777664816976) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q, 139777664816880) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v, 139777664816784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k, 139777664816928) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.weight, 139777659970608) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q, 139777664817024) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.weight, 139777664565792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out, 139777664816592) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0], 139777664816544) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1], 139777664816496) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj, 139777664816736) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj, 139777664816640) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj, 139777664816688) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out, 139777664816448) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k, 139777664816304) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.weight, 139777659970448) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q, 139777664816400) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.weight, 139777659970528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.processor, 139777664817168) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1, 139777664817600) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm, 139777664817456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu, 139777664817552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear, 139777664817504) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2, 139777664816256) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context, 139777664815920) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net, 139777664815776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0], 139777664815824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj, 139777664815728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1], 139777664815632) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2], 139777664815584) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context, 139777664817408) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm, 139777664817216) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu, 139777664817312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear, 139777664817264) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context, 139777664816208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4], 139777664817936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff, 139777664814288) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net, 139777664814048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0], 139777664814096) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj, 139777664814000) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1], 139777664813952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2], 139777664813904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn, 139777664815056) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k, 139777664814912) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q, 139777664814816) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v, 139777664814720) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k, 139777664814864) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.weight, 139777664570512) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q, 139777664814960) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.weight, 139777664571712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out, 139777664814528) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0], 139777664814480) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1], 139777664814432) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj, 139777664814672) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj, 139777664814576) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj, 139777664814624) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out, 139777664814384) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k, 139777664814240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.weight, 139777659971408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q, 139777664814336) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.weight, 139777664568832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.processor, 139777664815104) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1, 139777664815536) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm, 139777664815392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu, 139777664815488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear, 139777664815440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2, 139777664814192) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context, 139777664813856) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net, 139777664813712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0], 139777664813760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj, 139777664813664) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1], 139777664813568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2], 139777664813520) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context, 139777664815344) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm, 139777664815152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu, 139777664815248) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear, 139777664815200) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context, 139777664814144) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5], 139777664815872) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff, 139777664812224) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net, 139777664811984) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0], 139777664812032) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj, 139777664811936) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1], 139777664811888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2], 139777664811840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn, 139777664812992) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k, 139777664812848) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q, 139777664812752) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v, 139777664812656) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k, 139777664812800) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.weight, 139781198830368) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q, 139777664812896) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.weight, 139777664567072) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out, 139777664812464) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0], 139777664812416) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1], 139777664812368) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj, 139777664812608) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj, 139777664812512) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj, 139777664812560) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out, 139777664812320) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k, 139777664812176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.weight, 139781198834848) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q, 139777664812272) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.weight, 139777664184096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.processor, 139777664813040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1, 139777664813472) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm, 139777664813328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu, 139777664813424) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear, 139777664813376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2, 139777664812128) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context, 139777664811792) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net, 139777664811648) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0], 139777664811696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj, 139777664811600) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1], 139777664811504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2], 139777664811456) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context, 139777664813280) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm, 139777664813088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu, 139777664813184) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear, 139777664813136) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context, 139777664812080) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6], 139777664813808) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff, 139777674394928) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net, 139777674394688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0], 139777674394736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj, 139777674394640) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1], 139777674394592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2], 139777674394544) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn, 139777674383792) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k, 139777674395552) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q, 139777674395456) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v, 139777674395360) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k, 139777674395504) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.weight, 139777664177376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q, 139777674395600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.weight, 139777659972208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out, 139777674395168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0], 139777674395120) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1], 139777674395072) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj, 139777674395312) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj, 139777674395216) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj, 139777674395264) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out, 139777674395024) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k, 139777674394880) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.weight, 139777664570752) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q, 139777674394976) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.weight, 139777664177456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.processor, 139777674380624) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1, 139777664811408) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm, 139777664811264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu, 139777664811360) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear, 139777664811312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2, 139777674394832) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context, 139777674394496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net, 139777674394352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0], 139777674394400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj, 139777674394256) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1], 139777665008928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2], 139777665009312) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context, 139777664811216) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm, 139777674384080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu, 139777664811120) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear, 139777664811072) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context, 139777674394784) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7], 139777664811744) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff, 139777663769216) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net, 139777663770080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0], 139777663770656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj, 139777663770416) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1], 139777663769504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2], 139777663769648) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn, 139777665009504) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k, 139780515283440) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q, 139786872078544) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v, 139780515283008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k, 139786872288784) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.weight, 139777664175376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q, 139777676533120) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.weight, 139781198836928) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out, 139777663769456) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0], 139777663769408) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1], 139777663769360) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj, 139780515287040) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj, 139780515287088) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj, 139780515287136) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out, 139777663769312) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k, 139777663769120) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.weight, 139777660418256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q, 139777663769264) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.weight, 139777664178256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.processor, 139777665009072) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1, 139777665009360) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm, 139777665009456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu, 139777665009408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear, 139777665008976) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2, 139777663769552) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context, 139777663769744) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net, 139777663769600) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0], 139777663769696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj, 139777663770176) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1], 139777663770128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2], 139777663770320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context, 139777665009120) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm, 139777665007728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu, 139777665009024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear, 139777665007824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context, 139777663770752) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8], 139777674394448) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff, 139777663771472) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net, 139777663771712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0], 139777663771664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj, 139777663771760) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1], 139777663771808) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2], 139777663771856) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn, 139777663770848) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k, 139777663770800) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q, 139777663770944) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v, 139777663771040) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k, 139777663770896) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.weight, 139777659939520) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q, 139777663770704) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.weight, 139777664578912) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out, 139777663771232) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0], 139777663771280) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1], 139777663771328) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj, 139777663771088) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj, 139777663771184) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj, 139777663771136) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out, 139777663771376) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k, 139777663771520) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.weight, 139777660416816) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q, 139777663771424) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.weight, 139777664185616) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.processor, 139777663770512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1, 139777663769840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm, 139777663769984) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu, 139777663769888) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear, 139777663769936) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2, 139777663771568) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context, 139777663771904) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net, 139777663772048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0], 139777663772000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj, 139777663772096) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1], 139777663772192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2], 139777663772240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context, 139777663770560) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm, 139777663770032) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu, 139777663770224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear, 139777663770368) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context, 139777663771616) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9], 139777663769792) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff, 139777663773536) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net, 139777663773776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0], 139777663773728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj, 139777663773824) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1], 139777663773872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2], 139777663773920) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn, 139777663772768) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k, 139777663772912) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q, 139777663773008) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v, 139777663773104) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k, 139777663772960) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.weight, 139777664576272) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q, 139777663772864) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.weight, 139777664576352) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out, 139777663773296) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0], 139777663773344) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1], 139777663773392) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj, 139777663773152) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj, 139777663773248) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj, 139777663773200) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out, 139777663773440) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k, 139777663773584) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.weight, 139777664177936) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q, 139777663773488) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.weight, 139777664576992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.processor, 139777663772720) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1, 139777663772288) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm, 139777663772432) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu, 139777663772336) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear, 139777663772384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2, 139777663773632) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context, 139777663773968) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net, 139777663774112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0], 139777663774064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj, 139777663774160) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1], 139777663774256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2], 139777663774304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context, 139777663772480) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm, 139777663772672) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu, 139777663772576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear, 139777663772624) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context, 139777663773680) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10], 139777663771952) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff, 139777663775600) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net, 139777663775840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0], 139777663775792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj, 139777663775888) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1], 139777663775936) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2], 139777663775984) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn, 139777663774832) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k, 139777663774976) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q, 139777663775072) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v, 139777663775168) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k, 139777663775024) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.weight, 139777664181216) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q, 139777663774928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.weight, 139777664573792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out, 139777663775360) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0], 139777663775408) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1], 139777663775456) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj, 139777663775216) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj, 139777663775312) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj, 139777663775264) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out, 139777663775504) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k, 139777663775648) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.weight, 139777664181776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q, 139777663775552) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.weight, 139777664181136) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.processor, 139777663774784) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1, 139777663774352) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm, 139777663774496) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu, 139777663774400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear, 139777663774448) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2, 139777663775696) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context, 139777663776032) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net, 139777663776176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0], 139777663776128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj, 139777663776224) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1], 139777663776320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2], 139777663776368) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context, 139777663774544) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm, 139777663774736) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu, 139777663774640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear, 139777663774688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context, 139777663775744) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11], 139777663774016) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff, 139777663777664) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net, 139777663777904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0], 139777663777856) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj, 139777663777952) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1], 139777663778000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2], 139777663778048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn, 139777663776896) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k, 139777663777040) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q, 139777663777136) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v, 139777663777232) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k, 139777663777088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.weight, 139777664182176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q, 139777663776992) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.weight, 139777664571232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out, 139777663777424) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0], 139777663777472) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1], 139777663777520) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj, 139777663777280) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj, 139777663777376) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj, 139777663777328) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out, 139777663777568) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k, 139777663777712) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.weight, 139777664571952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q, 139777663777616) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.weight, 139777664178656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.processor, 139777663776848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1, 139777663776416) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm, 139777663776560) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu, 139777663776464) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear, 139777663776512) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2, 139777663777760) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context, 139777663778096) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net, 139777663778240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0], 139777663778192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj, 139777663778288) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1], 139777663778384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2], 139777663778432) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context, 139777663776608) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm, 139777663776800) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu, 139777663776704) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear, 139777663776752) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context, 139777663777808) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12], 139777663776080) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff, 139777663615952) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net, 139777663616240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0], 139777663616192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj, 139777663616288) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1], 139777663616336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2], 139777663616384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn, 139777663615184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k, 139777663615328) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q, 139777663615424) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v, 139777663615520) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k, 139777663615376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.weight, 139777664568592) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q, 139777663615280) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.weight, 139777664569312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out, 139777663615712) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0], 139777663615760) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1], 139777663615808) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj, 139777663615568) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj, 139777663615664) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj, 139777663615616) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out, 139777663615856) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k, 139777663616000) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.weight, 139777660406176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q, 139777663615904) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.weight, 139777664572832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.processor, 139777663615136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1, 139777663778480) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm, 139777663778624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu, 139777663778528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear, 139777663778576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2, 139777663616048) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context, 139777663616432) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net, 139777663616576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0], 139777663616528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj, 139777663616624) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1], 139777663616720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2], 139777663616768) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context, 139777663778672) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm, 139777663615088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu, 139777663778768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear, 139777663615040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context, 139777663616096) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13], 139777663778144) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff, 139777663618064) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net, 139777663618304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0], 139777663618256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj, 139777663618352) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1], 139777663618400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2], 139777663618448) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn, 139777663617296) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k, 139777663617440) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q, 139777663617536) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v, 139777663617632) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k, 139777663617488) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.weight, 139777659967648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q, 139777663617392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.weight, 139777659967728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out, 139777663617824) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0], 139777663617872) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1], 139777663617920) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj, 139777663617680) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj, 139777663617776) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj, 139777663617728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out, 139777663617968) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k, 139777663618112) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.weight, 139777664697824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q, 139777663618016) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.weight, 139777659967488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.processor, 139777663617248) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1, 139777663616816) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm, 139777663616960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu, 139777663616864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear, 139777663616912) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2, 139777663618160) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context, 139777663618496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net, 139777663618640) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0], 139777663618592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj, 139777663618688) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1], 139777663618784) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2], 139777663618832) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context, 139777663617008) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm, 139777663617200) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu, 139777663617104) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear, 139777663617152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context, 139777663618208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14], 139777663616480) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff, 139777663620128) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net, 139777663620368) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0], 139777663620320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj, 139777663620416) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1], 139777663620464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2], 139777663620512) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn, 139777663619360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k, 139777663619504) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q, 139777663619600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v, 139777663619696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k, 139777663619552) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.weight, 139777659969328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q, 139777663619456) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.weight, 139777659969408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out, 139777663619888) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0], 139777663619936) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1], 139777663619984) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj, 139777663619744) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj, 139777663619840) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj, 139777663619792) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out, 139777663620032) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k, 139777663620176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.weight, 139777659969168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q, 139777663620080) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.weight, 139777659969248) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.processor, 139777663619312) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1, 139777663618880) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm, 139777663619024) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu, 139777663618928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear, 139777663618976) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2, 139777663620224) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context, 139777663620560) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net, 139777663620704) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0], 139777663620656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj, 139777663620752) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1], 139777663620848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2], 139777663620896) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context, 139777663619072) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm, 139777663619264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu, 139777663619168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear, 139777663619216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context, 139777663620272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15], 139777663618544) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff, 139777663622192) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net, 139777663622480) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0], 139777663622384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj, 139777663622528) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1], 139777663622576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2], 139777663622624) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn, 139777663621424) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k, 139777663621568) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q, 139777663621664) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v, 139777663621760) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k, 139777663621616) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.weight, 139777664583696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q, 139777663621520) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.weight, 139777664581776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out, 139777663621952) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0], 139777663622000) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1], 139777663622048) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj, 139777663621808) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj, 139777663621904) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj, 139777663621856) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out, 139777663622096) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k, 139777663622240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.weight, 139777664574432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q, 139777663622144) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.weight, 139777664746656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.processor, 139777663621376) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1, 139777663620944) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm, 139777663621088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu, 139777663620992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear, 139777663621040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2, 139777663622288) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context, 139777663622672) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net, 139777663622816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0], 139777663622768) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj, 139777663622864) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1], 139777663622960) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2], 139777663623008) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context, 139777663621136) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm, 139777663621328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu, 139777663621232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear, 139777663621280) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context, 139777663622336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16], 139777663620608) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff, 139777663624352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net, 139777663624592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0], 139777663624544) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj, 139777663624640) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1], 139777663624688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2], 139777663624736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn, 139777663623584) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k, 139777663623728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q, 139777663623824) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v, 139777663623920) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k, 139777663623776) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.weight, 139777664327232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q, 139777663623680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.weight, 139777664594016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out, 139777663624112) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0], 139777663624160) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1], 139777663624208) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj, 139777663623968) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj, 139777663624064) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj, 139777663624016) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out, 139777663624256) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k, 139777663624400) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.weight, 139777664592896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q, 139777663624304) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.weight, 139777664595776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.processor, 139777663623536) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1, 139777663623056) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm, 139777663623248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu, 139777663623152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear, 139777663623200) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2, 139777663624448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context, 139777663624784) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net, 139777663624928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0], 139777663624880) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj, 139777663624976) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1], 139777663625072) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2], 139777663625120) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context, 139777663623296) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm, 139777663623488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu, 139777663623392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear, 139777663623440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context, 139777663624496) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17], 139777663622720) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff, 139777663626416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net, 139777663626656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0], 139777663626608) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj, 139777663626704) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1], 139777663626752) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2], 139777663626800) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn, 139777663625648) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k, 139777663625792) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q, 139777663625888) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v, 139777663625984) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k, 139777663625840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.weight, 139777675123168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q, 139777663625744) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.weight, 139777660095440) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out, 139777663626176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0], 139777663626224) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1], 139777663626272) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj, 139777663626032) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj, 139777663626128) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj, 139777663626080) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out, 139777663626320) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k, 139777663626464) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.weight, 139777664329392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q, 139777663626368) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.weight, 139777664585456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.processor, 139777663625600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1, 139777663625168) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm, 139777663625312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu, 139777663625216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear, 139777663625264) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2, 139777663626512) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context, 139777663626848) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net, 139777663626992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0], 139777663626944) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj, 139777663627040) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1], 139777663627136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2], 139777663627184) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context, 139777663625360) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm, 139777663625552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu, 139777663625456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear, 139777663625504) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context, 139777663626560) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18], 139777663624832) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff, 139777663628480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net, 139777663628720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0], 139777663628672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj, 139777663628768) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1], 139777663628816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2], 139777663628864) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn, 139777663627712) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k, 139777663627856) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q, 139777663627952) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v, 139777663628048) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k, 139777663627904) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.weight, 139777664592016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q, 139777663627808) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.weight, 139777664592096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out, 139777663628240) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0], 139777663628288) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1], 139777663628336) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj, 139777663628096) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.training, 139786895605728) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj, 139777663628192) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.training, 139786895605728) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj, 139777663628144) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.training, 139786895605728) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out, 139777663628384) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.training, 139786895605728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k, 139777663628528) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.weight, 139777664758736) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q, 139777663628432) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.weight, 139777664333392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.processor, 139777663627664) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1, 139777663627232) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm, 139777663627376) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu, 139777663627280) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear, 139777663627328) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2, 139777663628576) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context, 139777663628912) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net, 139777663629056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0], 139777663629008) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj, 139777663629104) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.training, 139786895605728) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1], 139777663629200) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2], 139777663629248) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context, 139777663627424) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm, 139777663627616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu, 139777663627520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear, 139777663627568) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context, 139777663628624) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=L['self'].single_transformer_blocks, accessed_by=DictGetItemGuardAccessor(single_transformer_blocks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks, 139777663628960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0], 139777663626896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn, 139777663629776) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k, 139777663629920) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q, 139777663630016) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v, 139777663630064) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k, 139777663629968) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.weight, 139777659813168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q, 139777663629824) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.weight, 139777660410496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.processor, 139777663629728) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm, 139777663629344) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm, 139777663629488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu, 139777663629392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear, 139777663629440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp, 139777663629632) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp, 139777663629584) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out, 139777663629680) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1], 139777663629296) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn, 139777663630544) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k, 139777663630688) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q, 139777663630784) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v, 139777663630832) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k, 139777663630736) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.weight, 139777660121344) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q, 139777663630592) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.weight, 139777664173856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.processor, 139777663630496) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm, 139777663630160) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm, 139777663630304) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu, 139777663630208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear, 139777663630256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp, 139777663630400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp, 139777663630352) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out, 139777663630448) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2], 139777663630112) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn, 139777663631312) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k, 139777662337184) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q, 139777662337280) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v, 139777662337328) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k, 139777662337232) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.weight, 139777660414656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q, 139777662337088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.weight, 139777664172096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.processor, 139777663631264) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm, 139777663630928) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm, 139777663631072) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu, 139777663630976) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear, 139777663631024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp, 139777663631168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp, 139777663631120) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out, 139777663631216) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3], 139777663630880) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn, 139777662337808) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k, 139777662337952) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q, 139777662338048) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v, 139777662338096) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k, 139777662338000) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.weight, 139777660434480) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q, 139777662337856) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.weight, 139777664183776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.processor, 139777662337760) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm, 139777662337424) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm, 139777662337568) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu, 139777662337472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear, 139777662337520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp, 139777662337664) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp, 139777662337616) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out, 139777662337712) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4], 139777662337376) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn, 139777662338576) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k, 139777662338720) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q, 139777662338816) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v, 139777662338864) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k, 139777662338768) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.weight, 139777664176976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q, 139777662338624) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.weight, 139777660379648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.processor, 139777662338528) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm, 139777662338192) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm, 139777662338336) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu, 139777662338240) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear, 139777662338288) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp, 139777662338432) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp, 139777662338384) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out, 139777662338480) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5], 139777662338144) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn, 139777662339344) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k, 139777662339488) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q, 139777662339584) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v, 139777662339632) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k, 139777662339536) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.weight, 139781198839008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q, 139777662339392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.weight, 139777664579472) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.processor, 139777662339296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm, 139777662338960) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm, 139777662339104) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu, 139777662339008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear, 139777662339056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp, 139777662339200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp, 139777662339152) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out, 139777662339248) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6], 139777662338912) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn, 139777662340112) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k, 139777662340256) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q, 139777662340352) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v, 139777662340400) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k, 139777662340304) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.weight, 139777664186496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q, 139777662340160) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.weight, 139777660410336) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.processor, 139777662340064) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm, 139777662339728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm, 139777662339872) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu, 139777662339776) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear, 139777662339824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp, 139777662339968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp, 139777662339920) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out, 139777662340016) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7], 139777662339680) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn, 139777662340880) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k, 139777662341024) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q, 139777662341120) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v, 139777662341168) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k, 139777662341072) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.weight, 139777664181696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q, 139777662340928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.weight, 139777664579392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.processor, 139777662340832) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm, 139777662340496) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm, 139777662340640) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu, 139777662340544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear, 139777662340592) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp, 139777662340736) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp, 139777662340688) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out, 139777662340784) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8], 139777662340448) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn, 139777662341648) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k, 139777662341792) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q, 139777662341888) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v, 139777662341936) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k, 139777662341840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.weight, 139777664477008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q, 139777662341696) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.weight, 139777664475008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.processor, 139777662341600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm, 139777662341264) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm, 139777662341408) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu, 139777662341312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear, 139777662341360) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp, 139777662341504) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp, 139777662341456) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out, 139777662341552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9], 139777662341216) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn, 139777662342416) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k, 139777662342560) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q, 139777662342656) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v, 139777662342704) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k, 139777662342608) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.weight, 139777664476128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q, 139777662342464) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.weight, 139777664479568) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.processor, 139777662342368) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm, 139777662342032) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm, 139777662342176) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu, 139777662342080) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear, 139777662342128) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp, 139777662342272) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp, 139777662342224) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out, 139777662342320) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10], 139777662341984) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn, 139777662343184) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k, 139777662343328) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q, 139777662343424) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v, 139777662343472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k, 139777662343376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.weight, 139777659800128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q, 139777662343232) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.weight, 139777664329632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.processor, 139777662343136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm, 139777662342800) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm, 139777662342944) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu, 139777662342848) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear, 139777662342896) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp, 139777662343040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp, 139777662342992) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out, 139777662343088) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11], 139777662342752) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn, 139777662343952) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k, 139777662344096) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q, 139777662344192) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v, 139777662344240) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k, 139777662344144) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.weight, 139777660418096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q, 139777662344000) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.weight, 139777659854400) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.processor, 139777662343904) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm, 139777662343568) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm, 139777662343712) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu, 139777662343616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear, 139777662343664) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp, 139777662343808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp, 139777662343760) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out, 139777662343856) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12], 139777662343520) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn, 139777662344720) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k, 139777662344864) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q, 139777662344960) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v, 139777662345008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k, 139777662344912) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.weight, 139777664575232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q, 139777662344768) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.weight, 139781198835088) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.processor, 139777662344672) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm, 139777662344336) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm, 139777662344480) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu, 139777662344384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear, 139777662344432) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp, 139777662344576) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp, 139777662344528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out, 139777662344624) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13], 139777662344288) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn, 139777662345488) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k, 139777662345632) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q, 139777662345728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v, 139777662345776) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k, 139777662345680) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.weight, 139777664581312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q, 139777662345536) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.weight, 139777659801808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.processor, 139777662345440) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm, 139777662345104) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm, 139777662345248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu, 139777662345152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear, 139777662345200) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp, 139777662345344) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp, 139777662345296) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out, 139777662345392) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14], 139777662345056) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn, 139777662346256) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k, 139777662346400) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q, 139777662346496) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v, 139777662346544) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k, 139777662346448) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.weight, 139777664751776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q, 139777662346304) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.weight, 139777664573952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.processor, 139777662346208) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm, 139777662345872) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm, 139777662346016) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu, 139777662345920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear, 139777662345968) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp, 139777662346112) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp, 139777662346064) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out, 139777662346160) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15], 139777662345824) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn, 139777662347024) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k, 139777662347168) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q, 139777662347264) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v, 139777662347312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k, 139777662347216) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.weight, 139777659855920) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q, 139777662347072) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.weight, 139777660118384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.processor, 139777662346976) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm, 139777662346640) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm, 139777662346784) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu, 139777662346688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear, 139777662346736) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp, 139777662346880) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp, 139777662346832) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out, 139777662346928) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16], 139777662346592) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn, 139777662347792) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k, 139777662347936) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q, 139777662348032) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v, 139777662348080) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k, 139777662347984) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.weight, 139777664323712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q, 139777662347840) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.weight, 139777664746576) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.processor, 139777662347744) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm, 139777662347408) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm, 139777662347552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu, 139777662347456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear, 139777662347504) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp, 139777662347648) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp, 139777662347600) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out, 139777662347696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17], 139777662347360) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn, 139777662348560) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k, 139777662348704) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q, 139777662348800) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v, 139777662348848) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k, 139777662348752) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.weight, 139777660124544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q, 139777662348608) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.weight, 139777659810208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.processor, 139777662348512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm, 139777662348176) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm, 139777662348320) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu, 139777662348224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear, 139777662348272) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp, 139777662348416) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp, 139777662348368) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out, 139777662348464) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18], 139777662348128) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn, 139777662349328) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k, 139777662349472) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q, 139777662349568) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v, 139777662349616) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k, 139777662349520) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.weight, 139777660373728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q, 139777662349376) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.weight, 139777660372208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.processor, 139777662349280) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm, 139777662348944) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm, 139777662349088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu, 139777662348992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear, 139777662349040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp, 139777662349184) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp, 139777662349136) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out, 139777662349232) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19], accessed_by=GetItemGuardAccessor(19) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19], 139777662348896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn, 139777662350096) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k, 139777662350240) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q, 139777662350336) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v, 139777662350384) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k, 139777662350288) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.weight, 139777664712304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q, 139777662350144) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.weight, 139777660377488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.processor, 139777662350048) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm, 139777662349712) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm, 139777662349856) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu, 139777662349760) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear, 139777662349808) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp, 139777662349952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp, 139777662349904) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out, 139777662350000) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20], accessed_by=GetItemGuardAccessor(20) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20], 139777662349664) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn, 139777662350864) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k, 139777662351008) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q, 139777662351104) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v, 139777662351152) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k, 139777662351056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.weight, 139777664182496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q, 139777662350912) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.weight, 139777664729392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.processor, 139777662350816) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm, 139777662350480) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm, 139777662350624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu, 139777662350528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear, 139777662350576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp, 139777662350720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp, 139777662350672) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out, 139777662350768) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21], accessed_by=GetItemGuardAccessor(21) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21], 139777662350432) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn, 139777662351632) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k, 139777662351776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q, 139777662351872) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v, 139777662351920) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k, 139777662351824) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.weight, 139777660377808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q, 139777662351680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.weight, 139777664725328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.processor, 139777662351584) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm, 139777662351248) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm, 139777662351392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu, 139777662351296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear, 139777662351344) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp, 139777662351488) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp, 139777662351440) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out, 139777662351536) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22], accessed_by=GetItemGuardAccessor(22) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22], 139777662351200) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn, 139777662352400) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k, 139777662352544) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q, 139777662352640) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v, 139777662352688) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k, 139777662352592) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.weight, 139777664757856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q, 139777662352448) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.weight, 139777660378528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.processor, 139777662352352) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm, 139777662352016) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm, 139777662352160) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu, 139777662352064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear, 139777662352112) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp, 139777662352256) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp, 139777662352208) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out, 139777662352304) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23], accessed_by=GetItemGuardAccessor(23) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23], 139777662351968) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn, 139777662353168) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k, 139777661026368) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q, 139777662353312) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v, 139777661026416) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k, 139777662353360) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.weight, 139777660445824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q, 139777662353216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.weight, 139777664332992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.processor, 139777662353120) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm, 139777662352784) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm, 139777662352928) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu, 139777662352832) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear, 139777662352880) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp, 139777662353024) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp, 139777662352976) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out, 139777662353072) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24], accessed_by=GetItemGuardAccessor(24) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24], 139777662352736) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn, 139777661026896) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k, 139777661027040) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q, 139777661027136) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v, 139777661027184) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k, 139777661027088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.weight, 139777660322976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q, 139777661026944) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.weight, 139777664578432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.processor, 139777661026848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm, 139777661026512) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm, 139777661026656) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu, 139777661026560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear, 139777661026608) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp, 139777661026752) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp, 139777661026704) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out, 139777661026800) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25], accessed_by=GetItemGuardAccessor(25) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25], 139777661026464) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn, 139777661027664) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k, 139777661027808) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q, 139777661027904) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v, 139777661027952) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k, 139777661027856) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.weight, 139777664182896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q, 139777661027712) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.weight, 139777664566032) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.processor, 139777661027616) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm, 139777661027280) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm, 139777661027424) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu, 139777661027328) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear, 139777661027376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp, 139777661027520) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp, 139777661027472) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out, 139777661027568) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26], accessed_by=GetItemGuardAccessor(26) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26], 139777661027232) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn, 139777661028432) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k, 139777661028576) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q, 139777661028672) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v, 139777661028720) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k, 139777661028624) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.weight, 139777661092384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q, 139777661028480) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.weight, 139777661093744) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.processor, 139777661028384) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm, 139777661028048) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm, 139777661028192) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu, 139777661028096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear, 139777661028144) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp, 139777661028288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp, 139777661028240) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out, 139777661028336) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27], accessed_by=GetItemGuardAccessor(27) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27], 139777661028000) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn, 139777661029200) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k, 139777661029344) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q, 139777661029440) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v, 139777661029488) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k, 139777661029392) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.weight, 139777661093504) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q, 139777661029248) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.weight, 139777661094864) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.processor, 139777661029152) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm, 139777661028816) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm, 139777661028960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu, 139777661028864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear, 139777661028912) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp, 139777661029056) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp, 139777661029008) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out, 139777661029104) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28], accessed_by=GetItemGuardAccessor(28) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28], 139777661028768) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn, 139777661030016) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k, 139777661030160) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q, 139777661030256) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v, 139777661030304) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k, 139777661030208) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.weight, 139777661094624) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q, 139777661030064) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.weight, 139777661096064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.processor, 139777661029968) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm, 139777661029584) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm, 139777661029728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu, 139777661029632) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear, 139777661029680) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp, 139777661029872) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp, 139777661029776) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out, 139777661029920) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29], accessed_by=GetItemGuardAccessor(29) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29], 139777661029536) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn, 139777661030784) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k, 139777661030928) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q, 139777661031024) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v, 139777661031072) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k, 139777661030976) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.weight, 139777661095824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q, 139777661030832) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.weight, 139777661097184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.processor, 139777661030736) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm, 139777661030400) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm, 139777661030544) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu, 139777661030448) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear, 139777661030496) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp, 139777661030640) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp, 139777661030592) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out, 139777661030688) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30], accessed_by=GetItemGuardAccessor(30) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30], 139777661030352) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn, 139777661031552) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k, 139777661031696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q, 139777661031792) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v, 139777661031840) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k, 139777661031744) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.weight, 139777661096944) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q, 139777661031600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.weight, 139777661098304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.processor, 139777661031504) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm, 139777661031168) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm, 139777661031312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu, 139777661031216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear, 139777661031264) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp, 139777661031408) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp, 139777661031360) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out, 139777661031456) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31], accessed_by=GetItemGuardAccessor(31) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31], 139777661031120) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn, 139777661032320) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k, 139777661032464) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q, 139777661032560) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v, 139777661032608) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k, 139777661032512) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.weight, 139777661098064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q, 139777661032368) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.weight, 139777661099424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.processor, 139777661032272) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm, 139777661031936) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm, 139777661032080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu, 139777661031984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear, 139777661032032) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp, 139777661032176) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp, 139777661032128) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out, 139777661032224) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32], accessed_by=GetItemGuardAccessor(32) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32], 139777661031888) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn, 139777661033088) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k, 139777661033232) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q, 139777661033328) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v, 139777661033376) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k, 139777661033280) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.weight, 139777661099184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q, 139777661033136) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.weight, 139777661100544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.processor, 139777661033040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm, 139777661032704) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm, 139777661032848) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu, 139777661032752) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear, 139777661032800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp, 139777661032944) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp, 139777661032896) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out, 139777661032992) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33], accessed_by=GetItemGuardAccessor(33) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33], 139777661032656) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn, 139777661033856) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k, 139777661034000) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q, 139777661034096) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v, 139777661034144) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k, 139777661034048) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.weight, 139777661100304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q, 139777661033904) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.weight, 139777661101664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.processor, 139777661033808) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm, 139777661033472) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm, 139777661033616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu, 139777661033520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear, 139777661033568) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp, 139777661033712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp, 139777661033664) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out, 139777661033760) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34], accessed_by=GetItemGuardAccessor(34) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34], 139777661033424) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn, 139777661034624) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k, 139777661034768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q, 139777661034864) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v, 139777661034912) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k, 139777661034816) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.weight, 139777661101424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q, 139777661034672) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.weight, 139777661102784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.processor, 139777661034576) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm, 139777661034240) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm, 139777661034384) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu, 139777661034288) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear, 139777661034336) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp, 139777661034480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp, 139777661034432) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out, 139777661034528) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35], accessed_by=GetItemGuardAccessor(35) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35], 139777661034192) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn, 139777661035392) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k, 139777661035536) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q, 139777661035632) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v, 139777661035680) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k, 139777661035584) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.weight, 139777661102544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q, 139777661035440) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.weight, 139777661103904) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.processor, 139777661035344) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm, 139777661035008) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm, 139777661035152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu, 139777661035056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear, 139777661035104) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp, 139777661035248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp, 139777661035200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out, 139777661035296) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36], accessed_by=GetItemGuardAccessor(36) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36], 139777661034960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn, 139777661036160) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k, 139777661036304) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q, 139777661036400) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v, 139777661036448) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k, 139777661036352) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.weight, 139777661103664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q, 139777661036208) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.weight, 139777661105024) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.processor, 139777661036112) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm, 139777661035776) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm, 139777661035920) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu, 139777661035824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear, 139777661035872) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp, 139777661036016) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp, 139777661035968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out, 139777661036064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37], accessed_by=GetItemGuardAccessor(37) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37], 139777661035728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn, 139777661036928) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k, 139777661037072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.training, 139786895605728) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q, 139777661037168) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.training, 139786895605728) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v, 139777661037216) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.training, 139786895605728) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k, 139777661037120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.weight, 139777661104784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q, 139777661036976) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.weight, 139777661106144) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.processor, 139777661036880) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm, 139777661036544) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm, 139777661036688) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu, 139777661036592) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear, 139777661036640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp, 139777661036784) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp, 139777661036736) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out, 139777661036832) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.training, 139786895605728) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['img_ids'], accessed_by=DictGetItemGuardAccessor(img_ids) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['img_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[4096, 3], stride=[3, 1]) # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['img_ids'], '_dynamo_dynamic_indices') == False # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['txt_ids'], accessed_by=DictGetItemGuardAccessor(txt_ids) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['txt_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[512, 3], stride=[3, 1]) # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['txt_ids'], '_dynamo_dynamic_indices') == False # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['guidance'], accessed_by=DictGetItemGuardAccessor(guidance) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['guidance'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.float32, device=0, requires_grad=False, size=[1], stride=[1]) # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['guidance'], '_dynamo_dynamic_indices') == False # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['timestep'], accessed_by=DictGetItemGuardAccessor(timestep) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['timestep'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1], stride=[1]) # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['timestep'], '_dynamo_dynamic_indices') == False # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['return_dict'], accessed_by=DictGetItemGuardAccessor(return_dict) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- ID_MATCH: ___check_obj_id(L['return_dict'], 139786895605728) # if not return_dict: # diffusers/src/diffusers/models/transformers/transformer_flux.py:555 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['hidden_states'], accessed_by=DictGetItemGuardAccessor(hidden_states) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 4096, 64], stride=[262144, 64, 1]) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['pooled_projections'], accessed_by=DictGetItemGuardAccessor(pooled_projections) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['pooled_projections'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 768], stride=[768, 1]) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['pooled_projections'], '_dynamo_dynamic_indices') == False # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['encoder_hidden_states'], accessed_by=DictGetItemGuardAccessor(encoder_hidden_states) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- TENSOR_MATCH: check_tensor(L['encoder_hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 512, 4096], stride=[2097152, 4096, 1]) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_HASATTR: hasattr(L['encoder_hidden_states'], '_dynamo_dynamic_indices') == False # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['joint_attention_kwargs'], accessed_by=DictGetItemGuardAccessor(joint_attention_kwargs) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- ID_MATCH: ___check_obj_id(L['joint_attention_kwargs'], 139786895690912) # if joint_attention_kwargs is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['controlnet_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_block_samples) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_block_samples'], 139786895690912) # if controlnet_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:502 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=L['controlnet_single_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_single_block_samples) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_single_block_samples'], 139786895690912) # if controlnet_single_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:538 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | +- GuardManager: source=G, accessed_by=GlobalsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['torch'], accessed_by=DictGetItemGuardAccessor(torch) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['torch'], 139786870267216) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['torch'].cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['torch'].cat, 139786867250080) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['torch'].float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- EQUALS_MATCH: G['torch'].float16 == torch.float16 # if encoder_hidden_states.dtype == torch.float16: # diffusers/src/diffusers/models/transformers/transformer_flux.py:200 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['USE_PEFT_BACKEND'], accessed_by=DictGetItemGuardAccessor(USE_PEFT_BACKEND) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['USE_PEFT_BACKEND'], 139786895605696) # if USE_PEFT_BACKEND: # diffusers/src/diffusers/models/transformers/transformer_flux.py:434 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['scale_lora_layers'], accessed_by=DictGetItemGuardAccessor(scale_lora_layers) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['scale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['scale_lora_layers'].__code__, 139780914343312) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['unscale_lora_layers'], accessed_by=DictGetItemGuardAccessor(unscale_lora_layers) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['unscale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['unscale_lora_layers'].__code__, 139780914343488) # unscale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:553 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__builtins_dict___0'], accessed_by=DictGetItemGuardAccessor(__builtins_dict___0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__builtins_dict___0']['int'], accessed_by=DictGetItemGuardAccessor(int) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___0']['int'], 139786895673536) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__builtins_dict___0']['len'], accessed_by=DictGetItemGuardAccessor(len) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___0']['len'], 139786873098768) # assert len(timesteps.shape) == 1, "Timesteps should be a 1d-array" # diffusers/src/diffusers/models/embeddings.py:54 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__builtins_dict___0']['set'], accessed_by=DictGetItemGuardAccessor(set) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___0']['set'], 139786895697184) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__builtins_dict___0']['range'], accessed_by=DictGetItemGuardAccessor(range) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___0']['range'], 139786895693664) # for i in range(n_axes): # diffusers/src/diffusers/models/embeddings.py:696 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__builtins_dict___0']['enumerate'], accessed_by=DictGetItemGuardAccessor(enumerate) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___0']['enumerate'], 139786895625344) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__builtins_dict___0']['isinstance'], accessed_by=DictGetItemGuardAccessor(isinstance) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___0']['isinstance'], 139786873098448) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention'], 139780514175456) # if len(args) > 0 or kwargs.get("scale", None) is not None: # diffusers/src/diffusers/models/attention.py:1197 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_embeddings) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'], 139780515771264) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np, accessed_by=GetAttrGuardAccessor(np) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np, 139786867275984) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, accessed_by=GetAttrGuardAccessor(ndarray) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, 139783921183872) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math, accessed_by=GetAttrGuardAccessor(math) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math, 139786870403888) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math.log, accessed_by=GetAttrGuardAccessor(log) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math.log, 139786870406768) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch, 139786870267216) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, 139786867250080) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, accessed_by=GetAttrGuardAccessor(cos) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, 139786867251280) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, accessed_by=GetAttrGuardAccessor(exp) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, 139786867252640) # emb = torch.exp(exponent) # diffusers/src/diffusers/models/embeddings.py:62 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, accessed_by=GetAttrGuardAccessor(sin) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, 139786867261040) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, accessed_by=GetAttrGuardAccessor(outer) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, 139786867273104) # freqs = torch.outer(pos, freqs) # type: ignore # [S, D/2] # diffusers/src/diffusers/models/embeddings.py:616 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, accessed_by=GetAttrGuardAccessor(stack) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, 139786867198048) # x_rotated = torch.stack([-x_imag, x_real], dim=-1).flatten(3) # diffusers/src/diffusers/models/embeddings.py:662 in apply_rotary_emb V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, accessed_by=GetAttrGuardAccessor(arange) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, 139786867122368) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float32 == torch.float32 # start=0, end=half_dim, dtype=torch.float32, device=timesteps.device # diffusers/src/diffusers/models/embeddings.py:58 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float64, accessed_by=GetAttrGuardAccessor(float64) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float64 == torch.float64 # freqs_dtype = torch.float32 if is_mps else torch.float64 # diffusers/src/diffusers/models/embeddings.py:695 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=GetAttrGuardAccessor(apply_rotary_emb) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, 139780515561872) # from .embeddings import apply_rotary_emb # diffusers/src/diffusers/models/attention_processor.py:1760 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], 139786895605696) # if use_real: # diffusers/src/diffusers/models/embeddings.py:653 in apply_rotary_emb V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1], accessed_by=GetItemGuardAccessor(1) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1] == -1 # if use_real_unbind_dim == -1: # diffusers/src/diffusers/models/embeddings.py:659 in apply_rotary_emb V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=GetAttrGuardAccessor(get_timestep_embedding) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, 139780515482768) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3] == 10000 # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=GetAttrGuardAccessor(get_1d_rotary_pos_embed) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, 139780515495616) # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=FuncDefaultsGuardAccessor V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0] == 10000.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2], accessed_by=GetItemGuardAccessor(2) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2] == 1.0 # 1.0 # diffusers/src/diffusers/models/embeddings.py:612 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3] == 1.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'], accessed_by=DictGetItemGuardAccessor(__import_torch_dot_nn_dot_modules_dot_module) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'], 139781214054384) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch, 139786870267216) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, accessed_by=GetAttrGuardAccessor(_C) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, 139786866637808) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, accessed_by=GetAttrGuardAccessor(_get_tracing_state) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, 139781219067088) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_pre_hooks) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_activations) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'], 139780515377728) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F, 139781213067664) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F.gelu, accessed_by=GetAttrGuardAccessor(gelu) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F.gelu, 139781219449200) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_normalization) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'], 139780513909392) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch, 139786870267216) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, accessed_by=GetAttrGuardAccessor(chunk) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, 139786867250240) # scale, shift = torch.chunk(emb, 2, dim=1) # diffusers/src/diffusers/models/normalization.py:305 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, accessed_by=GetAttrGuardAccessor(rsqrt) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, 139786867196688) # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float16 == torch.float16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float32 == torch.float32 # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16, accessed_by=GetAttrGuardAccessor(bfloat16) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16 == torch.bfloat16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention_processor) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'], 139780515379008) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F, 139781213067664) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, accessed_by=GetAttrGuardAccessor(scaled_dot_product_attention) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, 139781219601200) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch, 139786870267216) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, 139786867250080) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect, accessed_by=GetAttrGuardAccessor(inspect) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect, 139786871160896) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature, accessed_by=GetAttrGuardAccessor(signature) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, 139786868738336) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'], accessed_by=DictGetItemGuardAccessor(__import_peft_dot_tuners_dot_tuners_utils) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, accessed_by=GetAttrGuardAccessor(BaseTunerLayer) V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, 94650012074624) # from peft.tuners.tuners_utils import BaseTunerLayer # diffusers/src/diffusers/utils/peft_utils.py:113 in scale_lora_layers V0909 15:07:23.280000 139786888062784 torch/_dynamo/guards.py:2148] [0/0] [__guards] V0909 15:10:58.748000 139786888062784 torch/_dynamo/guards.py:2611] [0/1] [__recompiles] Recompiling function forward in /home/sayak/diffusers/src/diffusers/models/transformers/transformer_flux.py:388 V0909 15:10:58.748000 139786888062784 torch/_dynamo/guards.py:2611] [0/1] [__recompiles] triggered by the following guard failure(s): V0909 15:10:58.748000 139786888062784 torch/_dynamo/guards.py:2611] [0/1] [__recompiles] - ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139777664822256) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward W0909 15:13:21.360000 139786888062784 torch/fx/experimental/symbolic_shapes.py:4449] [0/1] xindex is not in var_ranges, defaulting to unknown range. V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2169] [0/1] [__guards] GUARDS: V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] TREE_GUARD_MANAGER: V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] +- RootGuardManager V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- DEFAULT_DEVICE: utils_device.CURRENT_DEVICE == None # _dynamo/output_graph.py:460 in init_ambient_guards V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GLOBAL_STATE: ___check_global_state() V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['self'], accessed_by=DictGetItemGuardAccessor(self) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- ID_MATCH: ___check_obj_id(L['self'], 139777664825136) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=L['self'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=L['self'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(L['self'].training, 139786895605728) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=L['self']._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].norm_out, accessed_by=DictGetItemGuardAccessor(norm_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out, 139777661036496) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].norm_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].norm_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.training, 139786895605728) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm, 139777661037408) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu, 139777661037312) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear, 139777661037360) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out, 139777661037264) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out.training, 139786895605728) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].pos_embed, accessed_by=DictGetItemGuardAccessor(pos_embed) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed, 139777664824800) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].pos_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].pos_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed.training, 139786895605728) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim, accessed_by=DictGetItemGuardAccessor(axes_dim) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].pos_embed.axes_dim, 139786895700800) # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- LENGTH_CHECK: len(L['self'].pos_embed.axes_dim) == 3 # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[0], accessed_by=TupleGetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[0] == 16 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[1], accessed_by=TupleGetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[1] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[2], accessed_by=TupleGetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[2] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].x_embedder, accessed_by=DictGetItemGuardAccessor(x_embedder) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder, 139777664824032) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].x_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].x_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder.training, 139786895605728) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].time_text_embed, accessed_by=DictGetItemGuardAccessor(time_text_embed) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed, 139777664824848) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].time_text_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.training, 139786895605728) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj, accessed_by=DictGetItemGuardAccessor(time_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj, 139777664824752) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.training, 139786895605728) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.scale, accessed_by=DictGetItemGuardAccessor(scale) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.scale == 1 # scale=self.scale, # diffusers/src/diffusers/models/embeddings.py:769 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.num_channels, accessed_by=DictGetItemGuardAccessor(num_channels) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.num_channels == 256 # self.num_channels, # diffusers/src/diffusers/models/embeddings.py:766 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.flip_sin_to_cos, accessed_by=DictGetItemGuardAccessor(flip_sin_to_cos) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.flip_sin_to_cos, 139786895605696) # flip_sin_to_cos=self.flip_sin_to_cos, # diffusers/src/diffusers/models/embeddings.py:767 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.downscale_freq_shift, accessed_by=DictGetItemGuardAccessor(downscale_freq_shift) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.downscale_freq_shift == 0 # downscale_freq_shift=self.downscale_freq_shift, # diffusers/src/diffusers/models/embeddings.py:768 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder, accessed_by=DictGetItemGuardAccessor(text_embedder) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder, 139777664824416) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.text_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.training, 139786895605728) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1, accessed_by=DictGetItemGuardAccessor(act_1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1, 139777664824128) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1.training, 139786895605728) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1, 139777664824176) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1.training, 139786895605728) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2, 139777664824080) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2.training, 139786895605728) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder, accessed_by=DictGetItemGuardAccessor(guidance_embedder) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder, 139777664824560) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.guidance_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.training, 139786895605728) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1, 139777664824512) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2, 139777664824464) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder, accessed_by=DictGetItemGuardAccessor(timestep_embedder) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder, 139777664824704) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.timestep_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.training, 139786895605728) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1, 139777664824656) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2, 139777664824608) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].context_embedder, accessed_by=DictGetItemGuardAccessor(context_embedder) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder, 139777664824320) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].context_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].context_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder.training, 139786895605728) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].transformer_blocks, accessed_by=DictGetItemGuardAccessor(transformer_blocks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks, 139777664823936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0], 139777664823984) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff, 139777664822544) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net, 139777664822304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0], 139777664822352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139728980164176) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A, 139728980157744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'], 139728980155728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['first'].weight, 139730414253968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B, 139728980152176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['first'], 139728980159952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.base_layer, 139777664822256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout, 139728980150064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['first'], 139728980153664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1], 139777664822208) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2], 139728980164416) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A, 139728979916928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['first'], 139728979916688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['first'].weight, 139730414245648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B, 139728979915584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B['first'], 139728979912368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].base_layer, 139777664822160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout, 139728979906896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout['first'], 139728980161824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn, 139777664823360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k, 139728979952400) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A, 139728979958400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['first'], 139728979959552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['first'].weight, 139730414870720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B, 139728979965600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B['first'], 139728979964640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.base_layer, 139777664823168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout, 139728979953168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout['first'], 139728979965360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q, 139728980350656) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A, 139728980357280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['first'], 139728979961040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['first'].weight, 139730397134992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B, 139728980353008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B['first'], 139728979952544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.base_layer, 139777664823072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout, 139728980358864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout['first'], 139728980360928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v, 139728979954416) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A, 139728979960560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['first'], 139728980322448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['first'].weight, 139730414877520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B, 139728979958352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B['first'], 139728980321296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.base_layer, 139777664822976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout, 139728979964784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout['first'], 139728979955040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k, 139777664823120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.weight, 139777659850240) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q, 139777664823216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.weight, 139777660412416) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out, 139777664822784) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0], 139728980312704) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A, 139728980320144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'], 139728978520976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['first'].weight, 139730414248368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B, 139728980322880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B['first'], 139728978513728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].base_layer, 139777664822736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout, 139728980320048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['first'], 139728980320240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1], 139777664822688) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj, 139728980321920) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A, 139728980321872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'], 139728980320960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['first'].weight, 139730414870560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B, 139728980322112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['first'], 139728980323216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.base_layer, 139777664822928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout, 139728980321680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['first'], 139728980322256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj, 139728980317456) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A, 139728980318704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'], 139728980313760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['first'].weight, 139730414249248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B, 139728980328400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['first'], 139728980320624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.base_layer, 139777664822832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout, 139728980318800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['first'], 139728980319376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj, 139728980327920) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A, 139728980320768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'], 139728980318368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['first'].weight, 139730414246208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B, 139728980317024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['first'], 139728980318416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.base_layer, 139777664822880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout, 139728980316016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['first'], 139728980321248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out, 139728978514016) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A, 139728978525632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'], 139728980150448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['first'].weight, 139730414245568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B, 139728980148576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B['first'], 139728980155248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.base_layer, 139777664822640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout, 139728978523904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['first'], 139728978525104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k, 139777664822496) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.weight, 139777664182976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q, 139777664822592) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.weight, 139777664182016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.processor, 139777664823408) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1, 139777664823840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm, 139777664823648) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu, 139777664823744) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear, 139728989471024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A, 139728989245152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['first'], 139728978841600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['first'].weight, 139728966287280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B, 139728966627440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B['first'], 139728978841936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.base_layer, 139777664823696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout, 139728989244624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout['first'], 139728989245536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear._active_adapter, 139786895671040) # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].norm1.linear._active_adapter) == 1 # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._active_adapter[0], accessed_by=ListGetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1.linear._active_adapter[0] == 'first' # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2, 139777664822448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context, 139777664822112) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net, 139777664821968) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0], 139777664822016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj, 139728979908240) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A, 139728979911984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'], 139728978645184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['first'].weight, 139730415927216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B, 139728979908624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['first'], 139728978645136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer, 139777664821920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout, 139728979903488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['first'], 139728979910016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1], 139777664821824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2], 139728978644320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A, 139728978645328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'], 139728978654640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['first'].weight, 139730415923936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B, 139728978644800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B['first'], 139728978654256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].base_layer, 139777664821776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout, 139728978644176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['first'], 139728978644944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context, 139777664823600) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm, 139777664823456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu, 139777664823552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear, 139728978846736) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A, 139728978842752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'], 139728978850336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['first'].weight, 139730389112592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B, 139728978844096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B['first'], 139728978839632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.base_layer, 139777664823504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout, 139728978850096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['first'], 139728978840160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context, 139777664822400) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1], 139777664823888) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff, 139777664820480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net, 139777664820240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0], 139777664820288) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj, 139728979049056) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A, 139728979048768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'], 139728979046464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['first'].weight, 139730415666176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B, 139728979036480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['first'], 139728978891712) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.base_layer, 139777664820192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout, 139728979036816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['first'], 139728979046320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1], 139777664820144) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2], 139728978900976) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A, 139728978901648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['first'], 139728978899104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['first'].weight, 139730415671936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B, 139728978902032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B['first'], 139728978892192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].base_layer, 139777664820096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout, 139728978892768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout['first'], 139728978901696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn, 139777664821248) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k, 139728978656416) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A, 139728978970128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['first'], 139728978969792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['first'].weight, 139730415823472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B, 139728978969408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B['first'], 139728978969600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.base_layer, 139777664821104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout, 139728978970272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout['first'], 139728978969504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q, 139728978650272) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A, 139728978649456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['first'], 139728978649888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['first'].weight, 139730415916576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B, 139728978643168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B['first'], 139728978656704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.base_layer, 139777664821008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout, 139728978650368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout['first'], 139728978650080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v, 139728978974208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A, 139728978972288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['first'], 139728978977136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['first'].weight, 139730415819632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B, 139728978972192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B['first'], 139728978983520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.base_layer, 139777664820912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout, 139728978975792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout['first'], 139728978974064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k, 139777664821056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.weight, 139777660113664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q, 139777664821152) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.weight, 139777659965328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out, 139777664820720) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0], 139728980484416) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A, 139728980488928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'], 139728980485328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['first'].weight, 139730415670336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B, 139728980483696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B['first'], 139728980490752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].base_layer, 139777664820672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout, 139728980488496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['first'], 139728980484272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1], 139777664820624) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj, 139728978977424) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A, 139728978798256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'], 139728978791248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['first'].weight, 139730415827392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B, 139728978795856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['first'], 139728978792016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.base_layer, 139777664820864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout, 139728978974592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['first'], 139728978983616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj, 139728980442464) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A, 139728980489888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'], 139728980484368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['first'].weight, 139730415818592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B, 139728980482064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['first'], 139728980484080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.base_layer, 139777664820768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout, 139728980490128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['first'], 139728980490224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj, 139728989471600) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A, 139728980438096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'], 139728980432528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['first'].weight, 139730415828352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B, 139728980442896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['first'], 139728980442512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.base_layer, 139777664820816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout, 139728980438720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['first'], 139728980429072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out, 139728980489408) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A, 139728980490896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'], 139728979036336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['first'].weight, 139730415678656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B, 139728980476928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B['first'], 139728979039648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.base_layer, 139777664820576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout, 139728980491664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['first'], 139728980490944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k, 139777664820432) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.weight, 139777659965408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q, 139777664820528) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.weight, 139777659965488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.processor, 139777664821296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1, 139777664821728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm, 139777664821584) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu, 139777664821680) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear, 139728978654064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A, 139728978653872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['first'], 139728978650032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['first'].weight, 139730415918416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B, 139728978652672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B['first'], 139728978649696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.base_layer, 139777664821632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout, 139728978654160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout['first'], 139728978654208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2, 139777664820384) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context, 139777664820048) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net, 139777664819904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0], 139777664819952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj, 139728978901744) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A, 139728978899200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'], 139728978897136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['first'].weight, 139730415676416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B, 139728978900880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['first'], 139728978898864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer, 139777664819856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout, 139728978902416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['first'], 139728978901936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1], 139777664819760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2], 139728978888448) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A, 139728978898384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'], 139728978891472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['first'].weight, 139730415488912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B, 139728978900064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B['first'], 139728978892528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].base_layer, 139777664819712) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout, 139728978901264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['first'], 139728978898912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context, 139777664821536) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm, 139777664821344) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu, 139777664821440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear, 139728978647152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A, 139728978647008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'], 139728978649312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['first'].weight, 139730415912576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B, 139728978644464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B['first'], 139728978642592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.base_layer, 139777664821392) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout, 139728978646624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['first'], 139728978647104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context, 139777664820336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2], 139777664822064) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff, 139777664818416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net, 139777664818176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0], 139777664818224) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj, 139728980519488) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A, 139728980519392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'], 139728980516800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['first'].weight, 139730415410112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B, 139728980519872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['first'], 139728980517760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.base_layer, 139777664818128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout, 139728980519536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['first'], 139728980518576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1], 139777664818080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2], 139728980516320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A, 139728980524576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['first'], 139728980521120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['first'].weight, 139730415110960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B, 139728980522800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B['first'], 139728980514640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].base_layer, 139777664818032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout, 139728980517472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout['first'], 139728980515264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn, 139777664819184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k, 139728977639600) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A, 139728977639744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['first'], 139728977634560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['first'].weight, 139730415496352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B, 139728977639696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B['first'], 139728977637296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.base_layer, 139777664819040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout, 139728977638784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout['first'], 139728977639312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q, 139728978789040) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A, 139728977625488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['first'], 139728977631968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['first'].weight, 139730415500032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B, 139728977625152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B['first'], 139728977638976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.base_layer, 139777664818944) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout, 139728977632256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout['first'], 139728977632112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v, 139728977632784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A, 139728977637488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['first'], 139728977632544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['first'].weight, 139730415487952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B, 139728977626256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B['first'], 139728977641328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.base_layer, 139777664818848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout, 139728977634896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout['first'], 139728977630816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k, 139777664818992) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.weight, 139777664567632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q, 139777664819088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.weight, 139777664569792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out, 139777664818656) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0], 139728979413168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A, 139728979426464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'], 139728979424688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['first'].weight, 139730415412832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B, 139728979426368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B['first'], 139728979413888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].base_layer, 139777664818608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout, 139728979422144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['first'], 139728979419312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1], 139777664818560) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj, 139728978452848) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A, 139728978454480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'], 139728978459568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['first'].weight, 139730415418672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B, 139728978456976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['first'], 139728978460096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.base_layer, 139777664818800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout, 139728978444448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['first'], 139728978445888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj, 139728978506368) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A, 139728978499120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'], 139728979418352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['first'].weight, 139730415414432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B, 139728978499360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['first'], 139728979412064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.base_layer, 139777664818704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout, 139728978500992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['first'], 139728978500896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj, 139728978458032) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A, 139728978458608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'], 139728978503008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['first'].weight, 139730415409232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B, 139728978456064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['first'], 139728978508192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.base_layer, 139777664818752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout, 139728978450208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['first'], 139728978445936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out, 139728979423344) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A, 139728979419120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'], 139728980516944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['first'].weight, 139730415410992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B, 139728979420656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B['first'], 139728980519680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.base_layer, 139777664818512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout, 139728979421040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['first'], 139728979425840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k, 139777664818368) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.weight, 139777664568672) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q, 139777664818464) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.weight, 139777664567552) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.processor, 139777664819232) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1, 139777664819664) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm, 139777664819520) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu, 139777664819616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear, 139728978887296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A, 139728978887584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['first'], 139728978896080) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['first'].weight, 139730415498352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B, 139728978887824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B['first'], 139728978897280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.base_layer, 139777664819568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout, 139728978897568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout['first'], 139728978896128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2, 139777664818320) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context, 139777664817984) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net, 139777664817840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0], 139777664817888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj, 139728980516272) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A, 139728980512000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'], 139728979093264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['first'].weight, 139730415117040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B, 139728979094896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['first'], 139728979095328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer, 139777664817792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout, 139728980522944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['first'], 139728980524816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1], 139777664817696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2], 139728979089472) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A, 139728979092208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'], 139728978751456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['first'].weight, 139730415114160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B, 139728979097440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B['first'], 139728978754384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].base_layer, 139777664817648) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout, 139728979097632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['first'], 139728979092976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context, 139777664819472) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm, 139777664819280) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu, 139777664819376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear, 139728978901024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A, 139728978900784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'], 139728978898768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['first'].weight, 139730415499712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B, 139728978900256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B['first'], 139728978898000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.base_layer, 139777664819328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout, 139728978888160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['first'], 139728978889648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context, 139777664818272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3], 139777664820000) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff, 139777664816352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net, 139777664816112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0], 139777664816160) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj, 139728977664448) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A, 139728977662576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'], 139728977672512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['first'].weight, 139730416951888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B, 139728977663920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['first'], 139728977669728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.base_layer, 139777664816064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout, 139728977665504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['first'], 139728977665072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1], 139777664816016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2], 139728977672992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A, 139728977665408) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['first'], 139728979344592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['first'].weight, 139730416944928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B, 139728977664592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B['first'], 139728979334416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].base_layer, 139777664815968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout, 139728977665168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout['first'], 139728977665552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn, 139777664817120) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k, 139728979281472) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A, 139728979293040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['first'], 139728979134640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['first'].weight, 139730417067216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B, 139728979293520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B['first'], 139728979139344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.base_layer, 139777664816976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout, 139728979280608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout['first'], 139728979291984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q, 139728979284784) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A, 139728979283584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['first'], 139728979295536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['first'].weight, 139730417060256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B, 139728979283536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B['first'], 139728979295152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.base_layer, 139777664816880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout, 139728979286608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout['first'], 139728979283968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v, 139728979144432) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A, 139728979137280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['first'], 139728979137904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['first'].weight, 139730417064416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B, 139728979137616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B['first'], 139728979141744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.base_layer, 139777664816784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout, 139728979136608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout['first'], 139728979136944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k, 139777664816928) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.weight, 139777659970608) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q, 139777664817024) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.weight, 139777664565792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out, 139777664816592) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0], 139728978811136) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A, 139728978811184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'], 139728978817520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['first'].weight, 139730416957888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B, 139728978817136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B['first'], 139728978814688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].base_layer, 139777664816544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout, 139728978820496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['first'], 139728978806480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1], 139777664816496) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj, 139728979138432) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A, 139728979135456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'], 139728979144816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['first'].weight, 139730417059376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B, 139728979138480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['first'], 139728979142560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.base_layer, 139777664816736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout, 139728979135216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['first'], 139728979134256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj, 139728979142224) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A, 139728978807968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'], 139728978820448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['first'].weight, 139730416958208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B, 139728978820160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['first'], 139728978813632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.base_layer, 139777664816640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout, 139728978819824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['first'], 139728978819968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj, 139728979142320) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A, 139728979133200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'], 139728979134064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['first'].weight, 139730417066976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B, 139728979136656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['first'], 139728979137376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.base_layer, 139777664816688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout, 139728979139776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['first'], 139728979137232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out, 139728978810896) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A, 139728978815984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'], 139728977663584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['first'].weight, 139730416949568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B, 139728978816224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B['first'], 139728977658736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.base_layer, 139777664816448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout, 139728978819920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['first'], 139728978809696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k, 139777664816304) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.weight, 139777659970448) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q, 139777664816400) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.weight, 139777659970528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.processor, 139777664817168) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1, 139777664817600) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm, 139777664817456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu, 139777664817552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear, 139728978751648) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A, 139728978754192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['first'], 139728979293760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['first'].weight, 139730415119360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B, 139728978746656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B['first'], 139728979286368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.base_layer, 139777664817504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout, 139728978751696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout['first'], 139728978748240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2, 139777664816256) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context, 139777664815920) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net, 139777664815776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0], 139777664815824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj, 139728979338592) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A, 139728979342672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'], 139728977454144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['first'].weight, 139730416946768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B, 139728979342000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['first'], 139728977450208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer, 139777664815728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout, 139728979339600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['first'], 139728979342576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1], 139777664815632) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2], 139728977447568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A, 139728979210256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'], 139728979206800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['first'].weight, 139730416762480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B, 139728979212656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B['first'], 139728979208576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].base_layer, 139777664815584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout, 139728977456496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['first'], 139728977456352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context, 139777664817408) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm, 139777664817216) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu, 139777664817312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear, 139728979286992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A, 139728979286800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'], 139728979286512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['first'].weight, 139730415114480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B, 139728979284832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B['first'], 139728979285168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.base_layer, 139777664817264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout, 139728979287088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['first'], 139728979287040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context, 139777664816208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4], 139777664817936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff, 139777664814288) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net, 139777664814048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0], 139777664814096) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj, 139728977554704) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A, 139728977558832) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'], 139728978048288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['first'].weight, 139730416639232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B, 139728978036816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['first'], 139728978044592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.base_layer, 139777664814000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout, 139728977558592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['first'], 139728977552640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1], 139777664813952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2], 139728978048528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A, 139728978046800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['first'], 139728978038352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['first'].weight, 139730416438144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B, 139728978035664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B['first'], 139728978038784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].base_layer, 139777664813904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout, 139728978037584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout['first'], 139728978037920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn, 139777664815056) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k, 139728978900544) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A, 139728977775632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['first'], 139728977776256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['first'].weight, 139730416754320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B, 139728977773328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B['first'], 139728977776304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.base_layer, 139777664814912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout, 139728977775248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout['first'], 139728977777744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q, 139728979210304) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A, 139728979375248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['first'], 139728979361904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['first'].weight, 139730416755760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B, 139728979362960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B['first'], 139728979370112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.base_layer, 139777664814816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout, 139728979373088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout['first'], 139728979366560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v, 139728977774192) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A, 139728977773136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['first'], 139728977774000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['first'].weight, 139730416752000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B, 139728977773184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B['first'], 139728977778368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.base_layer, 139777664814720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout, 139728977774528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout['first'], 139728977775488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k, 139777664814864) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.weight, 139777664570512) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q, 139777664814960) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.weight, 139777664571712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out, 139777664814528) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0], 139728979356560) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A, 139728979356416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'], 139728977552688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['first'].weight, 139730416644112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B, 139728979361408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B['first'], 139728977552496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].base_layer, 139777664814480) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout, 139728979348784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['first'], 139728979348976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1], 139777664814432) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj, 139728977779760) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A, 139728977788064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'], 139728977787008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['first'].weight, 139730416642912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B, 139728977780576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['first'], 139728977781488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.base_layer, 139777664814672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout, 139728977788784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['first'], 139728977779424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj, 139728980379200) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A, 139728980390816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'], 139728979359776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['first'].weight, 139730416642752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B, 139728980380976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['first'], 139728979346480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.base_layer, 139777664814576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout, 139728980381888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['first'], 139728980385872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj, 139728977773808) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A, 139728980384576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'], 139728980388704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['first'].weight, 139730416633232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B, 139728980389328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['first'], 139728980383184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.base_layer, 139777664814624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout, 139728980385968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['first'], 139728980385536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out, 139728977554176) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A, 139728977557488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'], 139728977543520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['first'].weight, 139730416643712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B, 139728977556720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B['first'], 139728977554368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.base_layer, 139777664814384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout, 139728977549616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['first'], 139728977549184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k, 139777664814240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.weight, 139777659971408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q, 139777664814336) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.weight, 139777664568832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.processor, 139777664815104) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1, 139777664815536) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm, 139777664815392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu, 139777664815488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear, 139728979208528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A, 139728979203584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['first'], 139728979205264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['first'].weight, 139730416762320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B, 139728979206608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B['first'], 139728979209296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.base_layer, 139777664815440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout, 139728979206416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout['first'], 139728979200896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2, 139777664814192) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context, 139777664813856) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net, 139777664813712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0], 139777664813760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj, 139728979365408) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A, 139728976457936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'], 139728976446944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['first'].weight, 139730416436544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B, 139728976456784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['first'], 139728976447088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer, 139777664813664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout, 139728976447712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['first'], 139728976451792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1], 139777664813568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2], 139728976450976) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A, 139728976460432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'], 139728977889888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['first'].weight, 139730416443264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B, 139728976451360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B['first'], 139728977894304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].base_layer, 139777664813520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout, 139728976451264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['first'], 139728976449056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context, 139777664815344) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm, 139777664815152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu, 139777664815248) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear, 139728979209344) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A, 139728979207856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'], 139728979199984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['first'].weight, 139730416748240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B, 139728979212560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B['first'], 139728979200320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.base_layer, 139777664815200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout, 139728979203008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['first'], 139728979200944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context, 139777664814144) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5], 139777664815872) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff, 139777664812224) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net, 139777664811984) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0], 139777664812032) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj, 139728978177056) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A, 139728979063520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'], 139728979060832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['first'].weight, 139730418132896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B, 139728979062752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['first'], 139728979059392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.base_layer, 139777664811936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout, 139728979054592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['first'], 139728979053920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1], 139777664811888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2], 139728979052672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A, 139728979058912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['first'], 139728978309856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['first'].weight, 139730418126016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B, 139728978307696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B['first'], 139728978312832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].base_layer, 139777664811840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout, 139728979052960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout['first'], 139728979063616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn, 139777664812992) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k, 139728977932304) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A, 139728977935808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['first'], 139728977934224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['first'].weight, 139730416306272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B, 139728977934464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B['first'], 139728977923280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.base_layer, 139777664812848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout, 139728977923232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout['first'], 139728977932400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q, 139728976378944) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A, 139728976373232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['first'], 139728977923520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['first'].weight, 139730416436624) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B, 139728976377984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B['first'], 139728977926112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.base_layer, 139777664812752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout, 139728976377840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout['first'], 139728976368864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v, 139728977920784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A, 139728977926160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['first'], 139728978181520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['first'].weight, 139730416304272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B, 139728978168224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B['first'], 139728978170288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.base_layer, 139777664812656) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout, 139728977931776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout['first'], 139728977927216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k, 139777664812800) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.weight, 139781198830368) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q, 139777664812896) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.weight, 139777664567072) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out, 139777664812464) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0], 139728978179600) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A, 139728978178160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'], 139728978169712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['first'].weight, 139730416308032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B, 139728978176816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B['first'], 139728978179504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].base_layer, 139777664812416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout, 139728978173360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['first'], 139728978166016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1], 139777664812368) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj, 139728978181328) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A, 139728978181280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'], 139728978180944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['first'].weight, 139730416319872) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B, 139728978180512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['first'], 139728978172544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.base_layer, 139777664812608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout, 139728978181376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['first'], 139728978179984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj, 139728978176864) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A, 139728978176480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'], 139728978174944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['first'].weight, 139730416313952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B, 139728978176048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['first'], 139728978173312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.base_layer, 139777664812512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout, 139728978173984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['first'], 139728978173216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj, 139728978175808) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A, 139728978174080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'], 139728978178832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['first'].weight, 139730416318672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B, 139728978169760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['first'], 139728978180416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.base_layer, 139777664812560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout, 139728978180032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['first'], 139728978179936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out, 139728978171728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A, 139728978171440) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'], 139728978177536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['first'].weight, 139730418129296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B, 139728978173648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B['first'], 139728978181664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.base_layer, 139777664812320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout, 139728978178016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['first'], 139728978177968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k, 139777664812176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.weight, 139781198834848) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q, 139777664812272) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.weight, 139777664184096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.processor, 139777664813040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1, 139777664813472) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm, 139777664813328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu, 139777664813424) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear, 139728977894256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A, 139728976365840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['first'], 139728976368240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['first'].weight, 139730416446144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B, 139728976367952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B['first'], 139728976367376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.base_layer, 139777664813376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout, 139728977894640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout['first'], 139728977896080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2, 139777664812128) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context, 139777664811792) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net, 139777664811648) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0], 139777664811696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj, 139728978312400) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A, 139728978302128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'], 139728976485808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['first'].weight, 139730418123136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B, 139728978306352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['first'], 139728976485136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer, 139777664811600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout, 139728978307552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['first'], 139728978299296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1], 139777664811504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2], 139728976484848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A, 139728976478656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'], 139728977987664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['first'].weight, 139730418137616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B, 139728976488352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B['first'], 139728977990592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].base_layer, 139777664811456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout, 139728976492384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['first'], 139728976480816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context, 139777664813280) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm, 139777664813088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu, 139777664813184) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear, 139728976368000) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A, 139728976366464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'], 139728976364928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['first'].weight, 139730416443504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B, 139728976366368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B['first'], 139728976364112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.base_layer, 139777664813136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout, 139728976367712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['first'], 139728976368096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context, 139777664812080) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6], 139777664813808) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff, 139777674394928) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net, 139777674394688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0], 139777674394736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj, 139728978388992) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A, 139728978390240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'], 139728978393312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['first'].weight, 139730417750864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B, 139728978390144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['first'], 139728978394800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.base_layer, 139777674394640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout, 139728978389664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['first'], 139728978389040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1], 139777674394592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2], 139728978395040) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A, 139728978381456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['first'], 139728978393024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['first'].weight, 139730417757984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B, 139728978385392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B['first'], 139728978385344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].base_layer, 139777674394544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout, 139728978392016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout['first'], 139728978392256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn, 139777674383792) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k, 139728977441888) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A, 139728977436224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['first'], 139728977960704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['first'].weight, 139730417925424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B, 139728977434160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B['first'], 139728977967088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.base_layer, 139777674395552) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout, 139728977443136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout['first'], 139728977444816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q, 139728978068288) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A, 139728978074768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['first'], 139728977435312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['first'].weight, 139730417922064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B, 139728978074912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B['first'], 139728977433920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.base_layer, 139777674395456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout, 139728978068624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout['first'], 139728978067904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v, 139728977959648) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A, 139728977958592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['first'], 139728977961040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['first'].weight, 139730417915104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B, 139728977959888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B['first'], 139728977964928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.base_layer, 139777674395360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout, 139728977964688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout['first'], 139728977959024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k, 139777674395504) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.weight, 139777664177376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q, 139777674395600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.weight, 139777659972208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out, 139777674395168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0], 139728977486000) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A, 139728977481536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'], 139728978378864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['first'].weight, 139730417749904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B, 139728978393744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B['first'], 139728978381984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].base_layer, 139777674395120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout, 139728977490512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['first'], 139728977485136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1], 139777674395072) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj, 139728977968144) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A, 139728977825792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'], 139728977837984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['first'].weight, 139730417925504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B, 139728977834000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['first'], 139728977825888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.base_layer, 139777674395312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout, 139728977960032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['first'], 139728977955280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj, 139728977485376) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A, 139728977485232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'], 139728977483552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['first'].weight, 139730417913664) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B, 139728977484944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['first'], 139728977490032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.base_layer, 139777674395216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout, 139728977491136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['first'], 139728977491568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj, 139728977834288) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A, 139728977478512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'], 139728977489792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['first'].weight, 139730417918544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B, 139728977488304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['first'], 139728977486192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.base_layer, 139777674395264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout, 139728977832896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['first'], 139728977829968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out, 139728978385440) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A, 139728978378960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'], 139728978379440) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['first'].weight, 139730417755904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B, 139728978381744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B['first'], 139728978381264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.base_layer, 139777674395024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout, 139728978385056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['first'], 139728978384576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k, 139777674394880) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.weight, 139777664570752) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q, 139777674394976) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.weight, 139777664177456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.processor, 139777674380624) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1, 139777664811408) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm, 139777664811264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu, 139777664811360) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear, 139728978080576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A, 139728978067760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['first'], 139728978071600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['first'].weight, 139730418128256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B, 139728978075152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B['first'], 139728978074816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.base_layer, 139777664811312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout, 139728978068144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout['first'], 139728978077072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2, 139777674394832) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context, 139777674394496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net, 139777674394352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0], 139777674394400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj, 139728978220448) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A, 139728978216320) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'], 139728978228896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['first'].weight, 139730417749504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B, 139728978215744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['first'], 139728978230528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer, 139777674394256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout, 139728978221168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['first'], 139728978220880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1], 139777665008928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2], 139728978230048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A, 139728978224336) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'], 139728978027200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['first'].weight, 139730417746864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B, 139728978025664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B['first'], 139728978020240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].base_layer, 139777665009312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout, 139728978230576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['first'], 139728978225776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context, 139777664811216) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm, 139777674384080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu, 139777664811120) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear, 139728978079040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A, 139728978078752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'], 139728978080048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['first'].weight, 139730418125216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B, 139728978083504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B['first'], 139728978076976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.base_layer, 139777664811072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout, 139728978067616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['first'], 139728978081920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context, 139777674394784) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7], 139777664811744) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff, 139777663769216) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net, 139777663770080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0], 139777663770656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj, 139728975703024) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A, 139728975697600) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'], 139728975705472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['first'].weight, 139730417418304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B, 139728975707776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['first'], 139728975705184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.base_layer, 139777663770416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout, 139728975695680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['first'], 139728975695728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1], 139777663769504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2], 139728975700336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A, 139728975701632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['first'], 139728975695776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['first'].weight, 139730417312576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B, 139728975702064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B['first'], 139728975693328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].base_layer, 139777663769648) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout, 139728975693472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout['first'], 139728975703888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn, 139777665009504) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k, 139728977295104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A, 139728976543904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['first'], 139728976546880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['first'].weight, 139730417545392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B, 139728976546928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B['first'], 139728976543856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.base_layer, 139780515283440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout, 139728977297264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout['first'], 139728977294864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q, 139728978033104) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A, 139728978025328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['first'], 139728977295632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['first'].weight, 139730417541872) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B, 139728978026288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B['first'], 139728977289152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.base_layer, 139786872078544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout, 139728978031472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout['first'], 139728978032816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v, 139728976546496) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A, 139728976556912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['first'], 139728976551728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['first'].weight, 139730417539952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B, 139728976551536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B['first'], 139728976546304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.base_layer, 139780515283008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout, 139728976550192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout['first'], 139728976545776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k, 139786872288784) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.weight, 139777664175376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q, 139777676533120) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.weight, 139781198836928) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out, 139777663769456) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0], 139728975827376) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A, 139728975834624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'], 139728975837408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['first'].weight, 139730417423824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B, 139728975835248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B['first'], 139728975837936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].base_layer, 139777663769408) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout, 139728975830448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['first'], 139728975834720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1], 139777663769360) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj, 139728975832032) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A, 139728975831552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'], 139728975837072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['first'].weight, 139730417431024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B, 139728975830928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['first'], 139728975828192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.base_layer, 139780515287040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout, 139728975830880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['first'], 139728975838272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj, 139728975835632) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A, 139728975835152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'], 139728975824496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['first'].weight, 139730417430384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B, 139728975833088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['first'], 139728975823920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.base_layer, 139780515287088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout, 139728975834144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['first'], 139728975831936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj, 139728975836496) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A, 139728975836448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'], 139728975837456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['first'].weight, 139730417424944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B, 139728975833856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['first'], 139728975839136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.base_layer, 139780515287136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout, 139728975831696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['first'], 139728975833664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out, 139728975831168) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A, 139728975704368) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'], 139728975698128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['first'].weight, 139730417426384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B, 139728975707680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B['first'], 139728975695632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.base_layer, 139777663769312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout, 139728975823824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['first'], 139728975831072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k, 139777663769120) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.weight, 139777660418256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q, 139777663769264) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.weight, 139777664178256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.processor, 139777665009072) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1, 139777665009360) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm, 139777665009456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu, 139777665009408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear, 139728978027824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A, 139728978034640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['first'], 139728978021680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['first'].weight, 139730417540752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B, 139728978026576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B['first'], 139728978019664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.base_layer, 139777665008976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout, 139728978025904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout['first'], 139728978030032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2, 139777663769552) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context, 139777663769744) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net, 139777663769600) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0], 139777663769696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj, 139728975693808) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A, 139728975698320) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'], 139728975702208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['first'].weight, 139730417316336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B, 139728975706336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['first'], 139728976637008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer, 139777663770176) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout, 139728975704656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['first'], 139728975693040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1], 139777663770128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2], 139728976639792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A, 139728976636240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'], 139728976641232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['first'].weight, 139730417319696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B, 139728976641472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B['first'], 139728976640896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].base_layer, 139777663770320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout, 139728976639360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['first'], 139728976632880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context, 139777665009120) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm, 139777665007728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu, 139777665009024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear, 139728978029936) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A, 139728978034064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'], 139728978032720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['first'].weight, 139730417535152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B, 139728978020000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B['first'], 139728978032528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.base_layer, 139777665007824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout, 139728978033776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['first'], 139728978029792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context, 139777663770752) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8], 139777674394448) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff, 139777663771472) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net, 139777663771712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0], 139777663771664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj, 139728975371312) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A, 139728975368768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'], 139728975369632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['first'].weight, 139730419046720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B, 139728975378800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['first'], 139728975369920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.base_layer, 139777663771760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout, 139728975374528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['first'], 139728975373040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1], 139777663771808) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2], 139728976625728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A, 139728975415328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['first'], 139728975427088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['first'].weight, 139730419051760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B, 139728975415040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B['first'], 139728975427520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].base_layer, 139777663771856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout, 139728975421328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout['first'], 139728975413744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn, 139777663770848) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k, 139728976642000) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A, 139728976637968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['first'], 139728976635184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['first'].weight, 139730417135232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B, 139728976636912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B['first'], 139728976635904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.base_layer, 139777663770800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout, 139728976627984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout['first'], 139728976640224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q, 139728976632832) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A, 139728976625872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['first'], 139728976632208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['first'].weight, 139730417303776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B, 139728976630720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B['first'], 139728976631776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.base_layer, 139777663770944) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout, 139728976637296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout['first'], 139728976639168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v, 139728978038448) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A, 139728975328112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['first'], 139728975325088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['first'].weight, 139730417138432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B, 139728975328256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B['first'], 139728975319520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.base_layer, 139777663771040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout, 139728975329216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout['first'], 139728975326144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k, 139777663770896) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.weight, 139777659939520) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q, 139777663770704) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.weight, 139777664578912) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out, 139777663771232) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0], 139728975372176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A, 139728975378704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'], 139728975369200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['first'].weight, 139730417130192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B, 139728975369728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B['first'], 139728975367424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].base_layer, 139777663771280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout, 139728975376160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['first'], 139728975369104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1], 139777663771328) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj, 139728975325568) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A, 139728976724656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'], 139728976726912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['first'].weight, 139730417126992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B, 139728976729312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['first'], 139728976730224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.base_layer, 139777663771088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout, 139728975325904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['first'], 139728975330128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj, 139728976726048) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A, 139728976737520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'], 139728976739680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['first'].weight, 139730417136192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B, 139728976737472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['first'], 139728976728640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.base_layer, 139777663771184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout, 139728976733056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['first'], 139728976731088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj, 139728976730704) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A, 139728976740064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'], 139728976733344) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['first'].weight, 139730417129232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B, 139728976740112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['first'], 139728976727488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.base_layer, 139777663771136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout, 139728976731040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['first'], 139728976731136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out, 139728975375248) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A, 139728975374192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'], 139728975366992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['first'].weight, 139730419048160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B, 139728975376016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B['first'], 139728975366656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.base_layer, 139777663771376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout, 139728975374048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['first'], 139728975376304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k, 139777663771520) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.weight, 139777660416816) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q, 139777663771424) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.weight, 139777664185616) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.processor, 139777663770512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1, 139777663769840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm, 139777663769984) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu, 139777663769888) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear, 139728976636384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A, 139728976638208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['first'], 139728976639504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['first'].weight, 139730417316896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B, 139728976638304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B['first'], 139728976638976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.base_layer, 139777663769936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout, 139728976636528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout['first'], 139728976637776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2, 139777663771568) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context, 139777663771904) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net, 139777663772048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0], 139777663772000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj, 139728975414992) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A, 139728975413936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'], 139728975991072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['first'].weight, 139730419046560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B, 139728975420080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['first'], 139728975998128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer, 139777663772096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout, 139728975429248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['first'], 139728975420752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1], 139777663772192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2], 139728976002496) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A, 139728976002592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'], 139728977408400) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['first'].weight, 139730419046880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B, 139728976002928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B['first'], 139728977404464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].base_layer, 139777663772240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout, 139728976002112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['first'], 139728976002208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context, 139777663770560) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm, 139777663770032) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu, 139777663770224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear, 139728976635376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A, 139728976639888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'], 139728976640176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['first'].weight, 139730417313136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B, 139728976640752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B['first'], 139728976640464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.base_layer, 139777663770368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout, 139728976627936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['first'], 139728976630432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context, 139777663771616) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9], 139777663769792) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff, 139777663773536) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net, 139777663773776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0], 139777663773728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj, 139728975643552) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A, 139728975649696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'], 139728975645856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['first'].weight, 139730418747888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B, 139728975643696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['first'], 139728975644272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.base_layer, 139777663773824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout, 139728975645328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['first'], 139728975645040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1], 139777663773872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2], 139728975646576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A, 139728975647392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['first'], 139728975653104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['first'].weight, 139730418746928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B, 139728975646000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B['first'], 139728975647536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].base_layer, 139777663773920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout, 139728975646096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout['first'], 139728975644320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn, 139777663772768) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k, 139728977399040) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A, 139728977397984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['first'], 139728977404416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['first'].weight, 139730418941552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B, 139728977397264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B['first'], 139728977404368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.base_layer, 139777663772912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout, 139728977396064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout['first'], 139728977406816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q, 139728977410176) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A, 139728977405760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['first'], 139728977401680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['first'].weight, 139730418927632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B, 139728977407248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B['first'], 139728977404560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.base_layer, 139777663773008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout, 139728977409408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout['first'], 139728977410128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v, 139728977397312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A, 139728977399904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['first'], 139728975558160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['first'].weight, 139730418941472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B, 139728977397552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B['first'], 139728975558256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.base_layer, 139777663773104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout, 139728977409648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout['first'], 139728977408976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k, 139777663772960) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.weight, 139777664576272) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q, 139777663772864) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.weight, 139777664576352) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out, 139777663773296) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0], 139728975767792) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A, 139728975606592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'], 139728975603376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['first'].weight, 139730418750368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B, 139728975600592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B['first'], 139728975597808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].base_layer, 139777663773344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout, 139728975597616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['first'], 139728975762416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1], 139777663773392) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj, 139728975547024) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A, 139728975556864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'], 139728975760112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['first'].weight, 139730418932112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B, 139728975549760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['first'], 139728975768416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.base_layer, 139777663773152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout, 139728975549280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['first'], 139728975546928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj, 139728975769184) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A, 139728975771920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'], 139728975763232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['first'].weight, 139730418750288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B, 139728975768608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['first'], 139728975770672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.base_layer, 139777663773248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout, 139728975757760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['first'], 139728975768560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj, 139728975769328) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A, 139728975759536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'], 139728975757712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['first'].weight, 139730418930832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B, 139728975767312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['first'], 139728975757808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.base_layer, 139777663773200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout, 139728975768464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['first'], 139728975768512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out, 139728975602704) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A, 139728975647584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'], 139728975644944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['first'].weight, 139730418761488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B, 139728975647488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B['first'], 139728975645904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.base_layer, 139777663773440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout, 139728975647728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['first'], 139728975646768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k, 139777663773584) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.weight, 139777664177936) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q, 139777663773488) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.weight, 139777664576992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.processor, 139777663772720) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1, 139777663772288) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm, 139777663772432) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu, 139777663772336) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear, 139728977408160) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A, 139728977408112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['first'], 139728977404848) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['first'].weight, 139730419042400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B, 139728977407968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B['first'], 139728977404944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.base_layer, 139777663772384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout, 139728977408208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout['first'], 139728977407584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2, 139777663773632) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context, 139777663773968) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net, 139777663774112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0], 139777663774064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj, 139728975646960) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A, 139728975924768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'], 139728975923472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['first'].weight, 139730418565504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B, 139728975936864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['first'], 139728975932688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer, 139777663774160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout, 139728975647680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['first'], 139728975647344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1], 139777663774256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2], 139728975897952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A, 139728975900880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'], 139728975891472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['first'].weight, 139730418578864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B, 139728975899680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B['first'], 139728975891616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].base_layer, 139777663774304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout, 139728975898720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['first'], 139728975898048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context, 139777663772480) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm, 139777663772672) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu, 139777663772576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear, 139728977403168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A, 139728977405184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'], 139728977406384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['first'].weight, 139730418929712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B, 139728977407392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B['first'], 139728977406240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.base_layer, 139777663772624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout, 139728977405232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['first'], 139728977405424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context, 139777663773680) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10], 139777663771952) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff, 139777663775600) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net, 139777663775840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0], 139777663775792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj, 139728976051504) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A, 139728976046176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'], 139728976050736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['first'].weight, 139730418280432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B, 139728976049488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['first'], 139728976040224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.base_layer, 139777663775888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout, 139728976037632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['first'], 139728976047856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1], 139777663775936) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2], 139728976049872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A, 139728974655024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['first'], 139728974657136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['first'].weight, 139730418284752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B, 139728974655600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B['first'], 139728974644560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].base_layer, 139777663775984) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout, 139728974658480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout['first'], 139728976037728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn, 139777663774832) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k, 139728976191376) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A, 139728976190992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['first'], 139728976193344) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['first'].weight, 139730418571904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B, 139728976192624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B['first'], 139728976192816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.base_layer, 139777663774976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout, 139728976191424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout['first'], 139728976191328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q, 139728975895216) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A, 139728975898432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['first'], 139728976183888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['first'].weight, 139730418574384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B, 139728976191280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B['first'], 139728976191520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.base_layer, 139777663775072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout, 139728975902272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout['first'], 139728975901312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v, 139728976199008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A, 139728976192576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['first'], 139728976190656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['first'].weight, 139730418566544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B, 139728976190896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B['first'], 139728976199056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.base_layer, 139777663775168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout, 139728976199152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout['first'], 139728976199584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k, 139777663775024) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.weight, 139777664181216) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q, 139777663774928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.weight, 139777664573792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out, 139777663775360) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0], 139728977387696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A, 139728977382080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'], 139728975944896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['first'].weight, 139730418462816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B, 139728977379872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B['first'], 139728975942736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].base_layer, 139777663775408) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout, 139728977380976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['first'], 139728977385536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1], 139777663775456) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj, 139728976188112) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A, 139728976199536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'], 139728976196512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['first'].weight, 139730418461616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B, 139728976197376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['first'], 139728976196320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.base_layer, 139777663775216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout, 139728976184032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['first'], 139728976188160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj, 139728975372992) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A, 139728976074912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'], 139728977379680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['first'].weight, 139730418466336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B, 139728976072368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['first'], 139728977389856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.base_layer, 139777663775312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout, 139728976081152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['first'], 139728976083600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj, 139728976188592) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A, 139728976184896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'], 139728976189264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['first'].weight, 139730418453376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B, 139728976190800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['first'], 139728976186384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.base_layer, 139777663775264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout, 139728976190416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['first'], 139728976188736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out, 139728975945472) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A, 139728975951952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'], 139728975938608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['first'].weight, 139730418452336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B, 139728975949888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B['first'], 139728975940048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.base_layer, 139777663775504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout, 139728975946672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['first'], 139728975945376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k, 139777663775648) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.weight, 139777664181776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q, 139777663775552) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.weight, 139777664181136) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.processor, 139777663774784) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1, 139777663774352) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm, 139777663774496) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu, 139777663774400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear, 139728975895840) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A, 139728975896704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['first'], 139728975888496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['first'].weight, 139730418574544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B, 139728975895888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B['first'], 139728975900208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.base_layer, 139777663774448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout, 139728975895600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout['first'], 139728975894400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2, 139777663775696) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context, 139777663776032) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net, 139777663776176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0], 139777663776128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj, 139728974653584) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A, 139728974644032) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'], 139728974643600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['first'].weight, 139730418277152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B, 139728974653632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['first'], 139728974643264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer, 139777663776224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout, 139728974651280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['first'], 139728974643408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1], 139777663776320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2], 139728974651664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A, 139728974657760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'], 139728974644128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['first'].weight, 139730418280032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B, 139728974647872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B['first'], 139728974646480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].base_layer, 139777663776368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout, 139728974645808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['first'], 139728974643936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context, 139777663774544) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm, 139777663774736) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu, 139777663774640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear, 139728975896224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A, 139728975898672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'], 139728975897136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['first'].weight, 139730418566464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B, 139728975898240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B['first'], 139728975896800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.base_layer, 139777663774688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout, 139728975896848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['first'], 139728975894112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context, 139777663775744) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11], 139777663774016) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff, 139777663777664) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net, 139777663777904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0], 139777663777856) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj, 139728974535152) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A, 139728974534960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'], 139728974532656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['first'].weight, 139732769966656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B, 139728974530784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['first'], 139728974533760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.base_layer, 139777663777952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout, 139728974536208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['first'], 139728974538032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1], 139777663778000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2], 139728974533472) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A, 139728974536640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['first'], 139728974483552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['first'].weight, 139732769980096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B, 139728974538512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B['first'], 139728974484512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].base_layer, 139777663778048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout, 139728974538080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout['first'], 139728974533520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn, 139777663776896) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k, 139728975482784) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A, 139728975482112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['first'], 139728975485808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['first'].weight, 139732770069840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B, 139728975483456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B['first'], 139728975485904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.base_layer, 139777663777040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout, 139728975482064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout['first'], 139728975482256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q, 139728977032192) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A, 139728977025616) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['first'], 139728977021584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['first'].weight, 139732770065120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B, 139728977026192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B['first'], 139728975481680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.base_layer, 139777663777136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout, 139728977032480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout['first'], 139728977033536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v, 139728975486240) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A, 139728975481776) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['first'], 139728975481536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['first'].weight, 139732770064000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B, 139728975484464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B['first'], 139728975482208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.base_layer, 139777663777232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout, 139728975485520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout['first'], 139728975486192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k, 139777663777088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.weight, 139777664182176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q, 139777663776992) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.weight, 139777664571232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out, 139777663777424) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0], 139728976310288) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A, 139728976308608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'], 139728974530064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['first'].weight, 139732770065040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B, 139728976310480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B['first'], 139728974535632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].base_layer, 139777663777472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout, 139728976301216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['first'], 139728976314224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1], 139777663777520) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj, 139728975480336) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A, 139728975495072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'], 139728976127712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['first'].weight, 139732770078000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B, 139728975494016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['first'], 139728976119792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.base_layer, 139777663777280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout, 139728975480528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['first'], 139728975480720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj, 139728977346128) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A, 139728977337728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'], 139728977330576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['first'].weight, 139732770072640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B, 139728977332448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['first'], 139728977332928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.base_layer, 139777663777376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout, 139728977344592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['first'], 139728977345072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj, 139728976133760) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A, 139728976122096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'], 139728977344256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['first'].weight, 139732770078560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B, 139728976122336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['first'], 139728977345600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.base_layer, 139777663777328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout, 139728976133616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['first'], 139728976133712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out, 139728974536352) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A, 139728974538176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'], 139728974539712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['first'].weight, 139732769965776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B, 139728974537120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B['first'], 139728974539760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.base_layer, 139777663777568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout, 139728974537264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['first'], 139728974536688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k, 139777663777712) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.weight, 139777664571952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q, 139777663777616) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.weight, 139777664178656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.processor, 139777663776848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1, 139777663776416) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm, 139777663776560) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu, 139777663776464) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear, 139728974655648) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A, 139728977035216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['first'], 139728977020912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['first'].weight, 139730418284032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B, 139728977021728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B['first'], 139728977028352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.base_layer, 139777663776512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout, 139728977022736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout['first'], 139728977025760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2, 139777663777760) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context, 139777663778096) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net, 139777663778240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0], 139777663778192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj, 139728974494400) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A, 139728974492960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'], 139728976328304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['first'].weight, 139732769971136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B, 139728974484848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['first'], 139728976329888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer, 139777663778288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout, 139728974490608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['first'], 139728974493488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1], 139777663778384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2], 139728976330320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A, 139728976322448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'], 139728976329456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['first'].weight, 139732769969216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B, 139728976318800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B['first'], 139728976320816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].base_layer, 139777663778432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout, 139728976322016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['first'], 139728976330608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context, 139777663776608) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm, 139777663776800) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu, 139777663776704) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear, 139728977020480) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A, 139728977028160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'], 139728977032624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['first'].weight, 139730418278832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B, 139728977034304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B['first'], 139728977021392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.base_layer, 139777663776752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout, 139728977020816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['first'], 139728977033200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context, 139777663777808) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12], 139777663776080) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff, 139777663615952) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net, 139777663616240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0], 139777663616192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj, 139728974844480) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A, 139728974849856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'], 139728974863168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['first'].weight, 139732769635392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B, 139728974850864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['first'], 139728974872192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.base_layer, 139777663616288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout, 139728974847936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['first'], 139728974849616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1], 139777663616336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2], 139728974741360) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A, 139728974736608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['first'], 139728974738912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['first'].weight, 139732769636272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B, 139728974739248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B['first'], 139728974738960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].base_layer, 139777663616384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout, 139728974738432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout['first'], 139728974739632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn, 139777663615184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k, 139728974553984) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A, 139728974556096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['first'], 139728973202432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['first'].weight, 139732769808416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B, 139728973206176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B['first'], 139728973209200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.base_layer, 139777663615328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout, 139728974559984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout['first'], 139728974553216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q, 139728974560128) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A, 139728974546784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['first'], 139728974558688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['first'].weight, 139732769806896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B, 139728974552976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B['first'], 139728974555616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.base_layer, 139777663615424) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout, 139728974551872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout['first'], 139728974554128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v, 139728973205312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A, 139728973207328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['first'], 139728973217744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['first'].weight, 139732769800416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B, 139728973205696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B['first'], 139728973217216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.base_layer, 139777663615520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout, 139728973214144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout['first'], 139728973207184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k, 139777663615376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.weight, 139777664568592) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q, 139777663615280) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.weight, 139777664569312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out, 139777663615712) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0], 139728974829920) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A, 139728974830064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'], 139728974832608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['first'].weight, 139732769625552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B, 139728974836784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B['first'], 139728975507616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].base_layer, 139777663615760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout, 139728974823536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['first'], 139728974829968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1], 139777663615808) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj, 139728973211984) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A, 139728973209824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'], 139728973217312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['first'].weight, 139732769812176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B, 139728973202960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['first'], 139728973215488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.base_layer, 139777663615568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout, 139728973216688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['first'], 139728973216736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj, 139728976184800) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A, 139728974830736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'], 139728974824544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['first'].weight, 139732769807296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B, 139728974832464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['first'], 139728974835824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.base_layer, 139777663615664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout, 139728974827232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['first'], 139728974831984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj, 139728973215776) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A, 139728973207664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'], 139728973208288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['first'].weight, 139732769814416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B, 139728973207520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['first'], 139728973203056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.base_layer, 139777663615616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout, 139728973202240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['first'], 139728973206992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out, 139728975500656) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A, 139728974855712) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'], 139728974855568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['first'].weight, 139732769632032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B, 139728974847168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B['first'], 139728974841936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.base_layer, 139777663615856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout, 139728975507472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['first'], 139728975500272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k, 139777663616000) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.weight, 139777660406176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q, 139777663615904) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.weight, 139777664572832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.processor, 139777663615136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1, 139777663778480) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm, 139777663778624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu, 139777663778528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear, 139728974328640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A, 139728974331376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['first'], 139728974330656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['first'].weight, 139732769967296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B, 139728974322304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B['first'], 139728974327104) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.base_layer, 139777663778576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout, 139728974325616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout['first'], 139728974328064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2, 139777663616048) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context, 139777663616432) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net, 139777663616576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0], 139777663616528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj, 139728974740688) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A, 139728974741264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'], 139728974737040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['first'].weight, 139732769631952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B, 139728974737568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['first'], 139728974735792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer, 139777663616624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout, 139728974735936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['first'], 139728974740640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1], 139777663616720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2], 139728974729456) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A, 139728974726960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'], 139728974728208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['first'].weight, 139732769625712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B, 139728974728784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B['first'], 139728974727776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].base_layer, 139777663616768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout, 139728974726384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['first'], 139728974725232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context, 139777663778672) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm, 139777663615088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu, 139777663778768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear, 139728974461008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A, 139728974456928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'], 139728974546064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['first'].weight, 139732769967376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B, 139728974451696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B['first'], 139728974553936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.base_layer, 139777663615040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout, 139728974460768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['first'], 139728974461056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context, 139777663616096) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13], 139777663778144) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff, 139777663618064) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net, 139777663618304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0], 139777663618256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj, 139728973424832) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A, 139728973418592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'], 139728973426560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['first'].weight, 139732769208608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B, 139728973425600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['first'], 139728973426608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.base_layer, 139777663618352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout, 139728973426080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['first'], 139728973429536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1], 139777663618400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2], 139728973423056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A, 139728973419792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['first'], 139728973429248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['first'].weight, 139732769201008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B, 139728973423200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B['first'], 139728973425312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].base_layer, 139777663618448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout, 139728973418352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout['first'], 139728973428912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn, 139777663617296) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k, 139728974585456) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A, 139728974587664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['first'], 139728974585120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['first'].weight, 139732769503360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B, 139728974591264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B['first'], 139728974587808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.base_layer, 139777663617440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout, 139728974592416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout['first'], 139728974586896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q, 139728974729072) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A, 139728974726192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['first'], 139728974588240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['first'].weight, 139732769489440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B, 139728974725280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B['first'], 139728974583296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.base_layer, 139777663617536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout, 139728974728400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout['first'], 139728974726816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v, 139728974579072) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A, 139728974582048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['first'], 139728974590640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['first'].weight, 139732769501120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B, 139728974591552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B['first'], 139728974584256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.base_layer, 139777663617632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout, 139728974592704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout['first'], 139728974588624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k, 139777663617488) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.weight, 139777659967648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q, 139777663617392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.weight, 139777659967728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out, 139777663617824) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0], 139728974382944) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A, 139728974390192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'], 139728974393792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['first'].weight, 139732769199328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B, 139728974391440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B['first'], 139728974390384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].base_layer, 139777663617872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout, 139728974381696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['first'], 139728974383184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1], 139777663617920) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj, 139728974593376) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A, 139728974578880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'], 139728974396432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['first'].weight, 139732769502320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B, 139728974587520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['first'], 139728974394704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.base_layer, 139777663617680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout, 139728974587280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['first'], 139728974593856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj, 139728974387312) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A, 139728974388752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'], 139728974393696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['first'].weight, 139732769202768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B, 139728974392160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['first'], 139728974389328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.base_layer, 139777663617776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout, 139728974394272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['first'], 139728974391296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj, 139728974384672) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A, 139728974382896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'], 139728974383040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['first'].weight, 139732769489040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B, 139728974381984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['first'], 139728974395424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.base_layer, 139777663617728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout, 139728974383376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['first'], 139728974383280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out, 139728974394800) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A, 139728974393072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'], 139728973417248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['first'].weight, 139732769204848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B, 139728974392400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B['first'], 139728973417776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.base_layer, 139777663617968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout, 139728974394368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['first'], 139728974383712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k, 139777663618112) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.weight, 139777664697824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q, 139777663618016) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.weight, 139777659967488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.processor, 139777663617248) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1, 139777663616816) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm, 139777663616960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu, 139777663616864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear, 139728974741024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A, 139728974738000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['first'], 139728974737376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['first'].weight, 139732769620672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B, 139728974729504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B['first'], 139728974737520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.base_layer, 139777663616912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout, 139728974741168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout['first'], 139728974733632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2, 139777663618160) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context, 139777663618496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net, 139777663618640) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0], 139777663618592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj, 139728973426896) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A, 139728973459904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'], 139728973458992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['first'].weight, 139732769199968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B, 139728973449824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['first'], 139728973456880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer, 139777663618688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout, 139728973428240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['first'], 139728973426656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1], 139777663618784) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2], 139728973462400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A, 139728973449008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'], 139728973450352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['first'].weight, 139732769063296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B, 139728973451744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B['first'], 139728973448624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].base_layer, 139777663618832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout, 139728973451936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['first'], 139728973451456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context, 139777663617008) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm, 139777663617200) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu, 139777663617104) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear, 139728974738816) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A, 139728974739152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'], 139728974726672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['first'].weight, 139732769491600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B, 139728974738624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B['first'], 139728974727200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.base_layer, 139777663617152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout, 139728974739392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['first'], 139728974738384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context, 139777663618208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14], 139777663616480) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff, 139777663620128) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net, 139777663620368) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0], 139777663620320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj, 139728975127808) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A, 139728975122960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'], 139728975124016) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['first'].weight, 139732770986704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B, 139728975122672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['first'], 139728975122528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.base_layer, 139777663620416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout, 139728975128480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['first'], 139728975127184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1], 139777663620464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2], 139728975120320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A, 139728975119072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['first'], 139728975132128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['first'].weight, 139732770980304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B, 139728975119648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B['first'], 139728975129392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].base_layer, 139777663620512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout, 139728975120416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout['first'], 139728975119888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn, 139777663619360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k, 139728975048192) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A, 139728975050592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['first'], 139728975045072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['first'].weight, 139732769066416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B, 139728975043248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B['first'], 139728975048672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.base_layer, 139777663619504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout, 139728975040080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout['first'], 139728975047376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q, 139728975049152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A, 139728975048720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['first'], 139728975048336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['first'].weight, 139732769065136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B, 139728975049008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B['first'], 139728975036576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.base_layer, 139777663619600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout, 139728975049200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout['first'], 139728975045456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v, 139728975044208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A, 139728975041568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['first'], 139728975047520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['first'].weight, 139732769069056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B, 139728975045648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B['first'], 139728975044976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.base_layer, 139777663619696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout, 139728975046464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout['first'], 139728975048768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k, 139777663619552) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.weight, 139777659969328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q, 139777663619456) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.weight, 139777659969408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out, 139777663619888) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0], 139728974689280) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A, 139728974684864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'], 139728974684480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['first'].weight, 139732770995504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B, 139728974692304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B['first'], 139728974682368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].base_layer, 139777663619936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout, 139728974688176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['first'], 139728974688944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1], 139777663619984) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj, 139728975046944) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A, 139728975044736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'], 139728974690432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['first'].weight, 139732770989744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B, 139728973391920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['first'], 139728974677856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.base_layer, 139777663619744) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout, 139728975050352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['first'], 139728975051120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj, 139728974688416) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A, 139728974686784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'], 139728974677184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['first'].weight, 139732770984544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B, 139728974681456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['first'], 139728974688080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.base_layer, 139777663619840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout, 139728974686208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['first'], 139728974686832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj, 139728974691680) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A, 139728974687024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'], 139728974680208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['first'].weight, 139732770991824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B, 139728974692208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['first'], 139728974686496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.base_layer, 139777663619792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout, 139728974690864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['first'], 139728974688800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out, 139728974682032) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A, 139728975127520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'], 139728975128960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['first'].weight, 139732770994384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B, 139728975129584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B['first'], 139728975128720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.base_layer, 139777663620032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout, 139728974689712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['first'], 139728974687312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k, 139777663620176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.weight, 139777659969168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q, 139777663620080) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.weight, 139777659969248) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.processor, 139777663619312) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1, 139777663618880) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm, 139777663619024) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu, 139777663618928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear, 139728973451408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A, 139728973450208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['first'], 139728973456784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['first'].weight, 139732769072496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B, 139728973456064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B['first'], 139728973461440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.base_layer, 139777663618976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout, 139728973451024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout['first'], 139728973452512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2, 139777663620224) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context, 139777663620560) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net, 139777663620704) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0], 139777663620656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj, 139728973207376) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A, 139728976288144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'], 139728976287616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['first'].weight, 139732770770992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B, 139728976288096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['first'], 139728976287328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer, 139777663620752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout, 139728976282912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['first'], 139728976283056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1], 139777663620848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2], 139728976282816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A, 139728976282768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'], 139728973308992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['first'].weight, 139732770771552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B, 139728976290016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B['first'], 139728973303472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].base_layer, 139777663620896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout, 139728976281808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['first'], 139728976287664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context, 139777663619072) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm, 139777663619264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu, 139777663619168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear, 139728973459232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A, 139728973452560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'], 139728975048816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['first'].weight, 139732769075936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B, 139728973450736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B['first'], 139728975049344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.base_layer, 139777663619216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout, 139728973451072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['first'], 139728973454288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context, 139777663620272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15], 139777663618544) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff, 139777663622192) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net, 139777663622480) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0], 139777663622384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj, 139728974814496) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A, 139728974813296) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'], 139728973578048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['first'].weight, 139732770420704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B, 139728973566528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['first'], 139728973564416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.base_layer, 139777663622528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout, 139728974812768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['first'], 139728974815216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1], 139777663622576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2], 139728973577136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A, 139728973566384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['first'], 139728973570800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['first'].weight, 139732770421504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B, 139728973563360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B['first'], 139728973568976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].base_layer, 139777663622624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout, 139728973578000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout['first'], 139728973576752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn, 139777663621424) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k, 139728973901744) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A, 139728973905200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['first'], 139728973893200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['first'].weight, 139732770609232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B, 139728973892144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B['first'], 139728973899056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.base_layer, 139777663621568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout, 139728973904864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout['first'], 139728973899488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q, 139728973890080) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A, 139728973891376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['first'], 139728973900544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['first'].weight, 139732770768752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B, 139728973905728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B['first'], 139728973893632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.base_layer, 139777663621664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout, 139728973904192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout['first'], 139728973905536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v, 139728974970256) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A, 139728974970832) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['first'], 139728974958304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['first'].weight, 139732770603472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B, 139728974964256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B['first'], 139728974965936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.base_layer, 139777663621760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout, 139728974964304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout['first'], 139728974970112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k, 139777663621616) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.weight, 139777664583696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q, 139777663621520) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.weight, 139777664581776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out, 139777663621952) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0], 139728973325616) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A, 139728973322592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'], 139728973325184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['first'].weight, 139732770411824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B, 139728973329792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B['first'], 139728973329648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].base_layer, 139777663622000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout, 139728973325712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['first'], 139728973329456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1], 139777663622048) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj, 139728974273008) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A, 139728974278576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'], 139728974276512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['first'].weight, 139732770618912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B, 139728974279008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['first'], 139728974275888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.base_layer, 139777663621808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout, 139728974279536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['first'], 139728974280640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj, 139728973329936) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A, 139728973328928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'], 139728973326864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['first'].weight, 139732770603792) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B, 139728973331136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['first'], 139728973327776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.base_layer, 139777663621904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout, 139728973329120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['first'], 139728973328880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj, 139728975132944) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A, 139728973318560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'], 139728973328544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['first'].weight, 139732770617232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B, 139728973319472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['first'], 139728973328784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.base_layer, 139777663621856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout, 139728975180704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['first'], 139728975183488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out, 139728973316352) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A, 139728973317120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'], 139728974820784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['first'].weight, 139732770407904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B, 139728973316496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B['first'], 139728974812000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.base_layer, 139777663622096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout, 139728973324752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['first'], 139728973316400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k, 139777663622240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.weight, 139777664574432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q, 139777663622144) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.weight, 139777664746656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.processor, 139777663621376) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1, 139777663620944) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm, 139777663621088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu, 139777663620992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear, 139728973306544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A, 139728973309520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['first'], 139728973309424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['first'].weight, 139732770778192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B, 139728973309088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B['first'], 139728973308656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.base_layer, 139777663621040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout, 139728973299968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout['first'], 139728973300352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2, 139777663622288) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context, 139777663622672) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net, 139777663622816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0], 139777663622768) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj, 139728973570896) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A, 139728973562208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'], 139728973570512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['first'].weight, 139732770419664) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B, 139728973570128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['first'], 139728973566480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer, 139777663622864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout, 139728973570944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['first'], 139728973570848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1], 139777663622960) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2], 139728973565616) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A, 139728973576128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'], 139728973565568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['first'].weight, 139732770413504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B, 139728973564272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B['first'], 139728973564944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].base_layer, 139777663623008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout, 139728973565808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['first'], 139728973577376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context, 139777663621136) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm, 139777663621328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu, 139777663621232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear, 139728973890800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A, 139728973905488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'], 139728973889696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['first'].weight, 139732770778032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B, 139728973890128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B['first'], 139728973897328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.base_layer, 139777663621280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout, 139728973905344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['first'], 139728973903376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context, 139777663622336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16], 139777663620608) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff, 139777663624352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net, 139777663624592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0], 139777663624544) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj, 139728972213568) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A, 139728974061024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'], 139728974054352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['first'].weight, 139732770190208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B, 139728974066448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['first'], 139728974062272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.base_layer, 139777663624640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout, 139728974068176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['first'], 139728974058144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1], 139777663624688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2], 139728974278192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A, 139728973822512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['first'], 139728975241536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['first'].weight, 139732770179408) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B, 139728973823472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B['first'], 139728975247248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].base_layer, 139777663624736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout, 139728973810224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout['first'], 139728973813536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn, 139777663623584) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k, 139728973512464) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A, 139728973502192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['first'], 139728973511504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['first'].weight, 139732770277872) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B, 139728973508960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B['first'], 139728973500560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.base_layer, 139777663623728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout, 139728973512224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout['first'], 139728973502096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q, 139728973510448) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A, 139728973499216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['first'], 139728973511696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['first'].weight, 139732770286352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B, 139728973500080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B['first'], 139728973512032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.base_layer, 139777663623824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout, 139728973500656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout['first'], 139728973506656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v, 139728972211408) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A, 139728972217504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['first'], 139728972216112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['first'].weight, 139732770285312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B, 139728972203872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B['first'], 139728972204016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.base_layer, 139777663623920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout, 139728972212512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout['first'], 139728972213472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k, 139777663623776) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.weight, 139777664327232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q, 139777663623680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.weight, 139777664594016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out, 139777663624112) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0], 139728972202096) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A, 139728972204352) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'], 139728972217168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['first'].weight, 139732770189568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B, 139728972203584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B['first'], 139728972216736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].base_layer, 139777663624160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout, 139728972204832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['first'], 139728972204928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1], 139777663624208) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj, 139728972209584) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A, 139728972204784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'], 139728972217984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['first'].weight, 139732770285552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B, 139728972205504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['first'], 139728972217840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.base_layer, 139777663623968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout, 139728972217264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['first'], 139728972217360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj, 139728972215824) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A, 139728972215200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'], 139728972205600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['first'].weight, 139732770181088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B, 139728972214912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['first'], 139728972205024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.base_layer, 139777663624064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout, 139728972216352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['first'], 139728972216304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj, 139728972216640) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A, 139728972218176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'], 139728972216208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['first'].weight, 139732770187168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B, 139728972217648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['first'], 139728972213424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.base_layer, 139777663624016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout, 139728972209872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['first'], 139728972214096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out, 139728972211792) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A, 139728972213952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'], 139728972203392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['first'].weight, 139732770193168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B, 139728972212464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B['first'], 139728972211024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.base_layer, 139777663624256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout, 139728972216400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['first'], 139728972216832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k, 139777663624400) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.weight, 139777664592896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q, 139777663624304) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.weight, 139777664595776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.processor, 139777663623536) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1, 139777663623056) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm, 139777663623248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu, 139777663623152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear, 139728973566000) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A, 139728973569504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['first'], 139728973562160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['first'].weight, 139732770283552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B, 139728973563840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B['first'], 139728973562064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.base_layer, 139777663623200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout, 139728973569216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout['first'], 139728973569120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2, 139777663624448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context, 139777663624784) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net, 139777663624928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0], 139777663624880) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj, 139728975241872) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A, 139728975233280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'], 139728975238032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['first'].weight, 139732770184688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B, 139728975249216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['first'], 139728972463408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer, 139777663624976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout, 139728975246912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['first'], 139728975247056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1], 139777663625072) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2], 139728972455344) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A, 139728972463312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'], 139728972463648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['first'].weight, 139732770177248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B, 139728972452704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B['first'], 139728972454288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].base_layer, 139777663625120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout, 139728972451552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['first'], 139728972455248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context, 139777663623296) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm, 139777663623488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu, 139777663623392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear, 139728973578096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A, 139728973506272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'], 139728973511600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['first'].weight, 139732770289792) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B, 139728973509152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B['first'], 139728973511312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.base_layer, 139777663623440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout, 139728973563456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['first'], 139728973573728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context, 139777663624496) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17], 139777663622720) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff, 139777663626416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net, 139777663626656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0], 139777663626608) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj, 139728974249120) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A, 139728974247968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'], 139728974247200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['first'].weight, 139732771895424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B, 139728974246144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['first'], 139728974246192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.base_layer, 139777663626704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout, 139728974248640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['first'], 139728974249216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1], 139777663626752) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2], 139728974246912) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A, 139728974246864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['first'], 139728974244560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['first'].weight, 139732771891584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B, 139728974245952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B['first'], 139728974243360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].base_layer, 139777663626800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout, 139728974247008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout['first'], 139728974246720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn, 139777663625648) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k, 139728972369440) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A, 139728972366224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['first'], 139728972374000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['first'].weight, 139732771996592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B, 139728972380480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B['first'], 139728973756880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.base_layer, 139777663625792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout, 139728972380288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout['first'], 139728972380336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q, 139728972371600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A, 139728972371024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['first'], 139728972379568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['first'].weight, 139732772011712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B, 139728972375680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B['first'], 139728972380048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.base_layer, 139777663625888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout, 139728972378896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout['first'], 139728972373712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v, 139728973750064) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A, 139728973742720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['first'], 139728975194448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['first'].weight, 139732771998352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B, 139728973750208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B['first'], 139728975193632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.base_layer, 139777663625984) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout, 139728973749632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout['first'], 139728973749728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k, 139777663625840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.weight, 139777675123168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q, 139777663625744) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.weight, 139777660095440) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out, 139777663626176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0], 139728974236016) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A, 139728974237456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'], 139728974243072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['first'].weight, 139732771883184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B, 139728974236976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B['first'], 139728974235008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].base_layer, 139777663626224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout, 139728974236256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['first'], 139728974235728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1], 139777663626272) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj, 139728975195984) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A, 139728975194880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'], 139728975192336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['first'].weight, 139732771890864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B, 139728975195696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['first'], 139728975196128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.base_layer, 139777663626032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout, 139728975192624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['first'], 139728975192768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj, 139728973716768) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A, 139728974235824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'], 139728974237168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['first'].weight, 139732771894944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B, 139728974236880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['first'], 139728974246000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.base_layer, 139777663626128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout, 139728974236496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['first'], 139728973714800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj, 139728973722000) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A, 139728973722240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'], 139728973717632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['first'].weight, 139732771896624) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B, 139728973714944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['first'], 139728973714512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.base_layer, 139777663626080) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout, 139728973721904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['first'], 139728973710192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out, 139728974237408) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A, 139728974247680) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'], 139728974244896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['first'].weight, 139732771895104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B, 139728974244848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B['first'], 139728974240912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.base_layer, 139777663626320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout, 139728974242208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['first'], 139728974234768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k, 139777663626464) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.weight, 139777664329392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q, 139777663626368) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.weight, 139777664585456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.processor, 139777663625600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1, 139777663625168) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm, 139777663625312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu, 139777663625216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear, 139728972458800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A, 139728972463744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['first'], 139728972373088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['first'].weight, 139732772003472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B, 139728972374240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B['first'], 139728972368000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.base_layer, 139777663625264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout, 139728972453472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout['first'], 139728972455056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2, 139777663626512) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context, 139777663626848) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net, 139777663626992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0], 139777663626944) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj, 139728974234960) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A, 139728974237216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'], 139728974243840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['first'].weight, 139732771707680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B, 139728974235584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['first'], 139728974243744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer, 139777663627040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout, 139728974235680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['first'], 139728974236544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1], 139777663627136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2], 139728974062464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A, 139728975001632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'], 139728974993520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['first'].weight, 139732771707920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B, 139728974994192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B['first'], 139728974993664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].base_layer, 139777663627184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout, 139728975002064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['first'], 139728974998368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context, 139777663625360) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm, 139777663625552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu, 139777663625456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear, 139728972375728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A, 139728972376304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'], 139728972375776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['first'].weight, 139732772005952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B, 139728972372272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B['first'], 139728972368864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.base_layer, 139777663625504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout, 139728972371792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['first'], 139728972375632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context, 139777663626560) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18], 139777663624832) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff, 139777663628480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net, 139777663628720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0], 139777663628672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj, 139728972775184) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A, 139728972762416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'], 139728972769040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['first'].weight, 139732771434432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B, 139728972762320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['first'], 139728972761504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.base_layer, 139777663628768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout, 139728972760784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['first'], 139728972766736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1], 139777663628816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2], 139728974236064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A, 139728972688512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['first'], 139728972692592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['first'].weight, 139732771436912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B, 139728972691776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B['first'], 139728972682416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].base_layer, 139777663628864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout, 139728972682800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout['first'], 139728972688944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn, 139777663627712) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k, 139728973977088) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A, 139728973976896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['first'], 139728974183008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['first'].weight, 139732771703600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B, 139728973975408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B['first'], 139728974182096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.base_layer, 139777663627856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout, 139728973977136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout['first'], 139728973976416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q, 139728973973920) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A, 139728973979296) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['first'], 139728973977616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['first'].weight, 139732771710960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B, 139728973976464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B['first'], 139728973977472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.base_layer, 139777663627952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout, 139728973972816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout['first'], 139728973973872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v, 139728974183152) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A, 139728974168848) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['first'], 139728974179408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['first'].weight, 139732771557024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B, 139728974184256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B['first'], 139728974168752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.base_layer, 139777663628048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout, 139728974182144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout['first'], 139728974180512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k, 139777663627904) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.weight, 139777664592016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q, 139777663627808) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.weight, 139777664592096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out, 139777663628240) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0], 139728972766976) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A, 139728972761744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'], 139728972760592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['first'].weight, 139732771557584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B, 139728972767168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B['first'], 139728972759104) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].base_layer, 139777663628288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout, 139728972774128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['first'], 139728972761888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_out[0].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_out[0].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_out[0].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1], 139777663628336) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj, 139728974183296) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A, 139728974182672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'], 139728974173840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['first'].weight, 139732771554144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B, 139728974169088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['first'], 139728974177056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.base_layer, 139777663628096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout, 139728974181616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['first'], 139728974173792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_k_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_k_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj, 139728972766208) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A, 139728972767552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'], 139728972767360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['first'].weight, 139732771555344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B, 139728972767600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['first'], 139728972770336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.base_layer, 139777663628192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout, 139728972766592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['first'], 139728972767984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_q_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_q_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj, 139728974173072) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A, 139728974173504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'], 139728972766640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['first'].weight, 139732771566224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B, 139728972774080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['first'], 139728972766544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.base_layer, 139777663628144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout, 139728974178928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['first'], 139728974179456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_v_proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_v_proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out, 139728972767888) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A, 139728972760208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'], 139728972768368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['first'].weight, 139732771431312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B, 139728972768608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B['first'], 139728972766496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.base_layer, 139777663628384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout, 139728972760544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['first'], 139728972759488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_add_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_add_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_add_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k, 139777663628528) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.weight, 139777664758736) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q, 139777663628432) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.weight, 139777664333392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.processor, 139777663627664) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1, 139777663627232) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm, 139777663627376) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu, 139777663627280) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear, 139728974995488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A, 139728973972048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['first'], 139728973973536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['first'].weight, 139732771701920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B, 139728973971904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B['first'], 139728973973344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.base_layer, 139777663627328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout, 139728973978864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout['first'], 139728974993568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].norm1.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2, 139777663628576) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context, 139777663628912) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net, 139777663629056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0], 139777663629008) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj, 139728972759776) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A, 139728972501184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'], 139728972508720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['first'].weight, 139732771431792) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B, 139728972510736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['first'], 139728972502096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer, 139777663629104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout, 139728972512896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['first'], 139728972512992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1], 139777663629200) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2], 139728972498400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A, 139728972505792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'], 139728972497296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['first'].weight, 139732771436432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B, 139728972500464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B['first'], 139728972497344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].base_layer, 139777663629248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout, 139728972510160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['first'], 139728972498448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[2].scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[2].scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[2].use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context, 139777663627424) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm, 139777663627616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu, 139777663627520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear, 139728973986400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A, 139728973986208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'], 139728973984912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['first'].weight, 139732771715440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B, 139728973986160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B['first'], 139728973978816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.base_layer, 139777663627568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout, 139728973986688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['first'], 139728973986544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1_context.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].norm1_context.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1_context.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context, 139777663628624) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=L['self'].single_transformer_blocks, accessed_by=DictGetItemGuardAccessor(single_transformer_blocks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks, 139777663628960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0], 139777663626896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn, 139777663629776) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k, 139728973631584) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A, 139728973630048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'], 139728973627600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['first'].weight, 139732771324288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B, 139728973628368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B['first'], 139728973627456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.base_layer, 139777663629920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout, 139728973631632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['first'], 139728973631008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q, 139728975203488) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A, 139728975210112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'], 139728973629376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['first'].weight, 139732771326608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B, 139728975214384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B['first'], 139728973629280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.base_layer, 139777663630016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout, 139728975203872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['first'], 139728975216544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v, 139728973627696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A, 139728973629568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'], 139728973628944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['first'].weight, 139732771336048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B, 139728973629664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B['first'], 139728973639552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.base_layer, 139777663630064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout, 139728973628176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['first'], 139728973627504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k, 139777663629968) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.weight, 139777659813168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q, 139777663629824) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.weight, 139777660410496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.processor, 139777663629728) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm, 139777663629344) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm, 139777663629488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu, 139777663629392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear, 139728972497728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A, 139728972498064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'], 139728972498160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['first'].weight, 139732771431232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B, 139728972498736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B['first'], 139728972498688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.base_layer, 139777663629440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout, 139728972497248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['first'], 139728972497056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp, 139777663629632) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp, 139728972501952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A, 139728972504400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'], 139728972512608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['first'].weight, 139732771333808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B, 139728972512464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B['first'], 139728975201280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.base_layer, 139777663629584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout, 139728972510640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['first'], 139728972504688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out, 139728975204064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A, 139728975202480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['first'], 139728975214048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['first'].weight, 139732771327168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B, 139728975204304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B['first'], 139728975214864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.base_layer, 139777663629680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout, 139728975202432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout['first'], 139728975203728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1], 139777663629296) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn, 139777663630544) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k, 139728974145968) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A, 139728974141264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'], 139728974100416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['first'].weight, 139732771183152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B, 139728974093744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B['first'], 139728974098688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.base_layer, 139777663630688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout, 139728974147696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['first'], 139728974140832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q, 139728973400240) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A, 139728973410608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'], 139728974137184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['first'].weight, 139732771187232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B, 139728974137520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B['first'], 139728974135360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.base_layer, 139777663630784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout, 139728973400336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['first'], 139728973411424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v, 139728974100320) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A, 139728974092016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'], 139728972335904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['first'].weight, 139732771191472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B, 139728974102240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B['first'], 139728972344400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.base_layer, 139777663630832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout, 139728974100608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['first'], 139728974102432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k, 139777663630736) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.weight, 139777660121344) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q, 139777663630592) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.weight, 139777664173856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.processor, 139777663630496) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm, 139777663630160) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm, 139777663630304) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu, 139777663630208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear, 139728973632112) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A, 139728973630864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'], 139728973862256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['first'].weight, 139732771333088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B, 139728973632160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B['first'], 139728973859664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.base_layer, 139777663630256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout, 139728973643392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['first'], 139728973630576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp, 139777663630400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp, 139728973863936) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A, 139728973866144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'], 139728973412816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['first'].weight, 139732771327248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B, 139728973863024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B['first'], 139728973404176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.base_layer, 139777663630352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout, 139728973859856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['first'], 139728973862880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out, 139728973410368) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A, 139728973401008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['first'], 139728973408496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['first'].weight, 139732771177792) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B, 139728973399616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B['first'], 139728973400432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.base_layer, 139777663630448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout, 139728973406768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout['first'], 139728973410704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2], 139777663630112) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn, 139777663631312) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k, 139728972985248) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A, 139728972986688) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'], 139728972983904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['first'].weight, 139732773125184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B, 139728972986976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B['first'], 139728972985440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.base_layer, 139777662337184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout, 139728972979296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['first'], 139728972987120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q, 139728972288048) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A, 139728972295200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'], 139728972974928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['first'].weight, 139732773122704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B, 139728972290448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B['first'], 139728972980112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.base_layer, 139777662337280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout, 139728972292320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['first'], 139728972291024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v, 139728972974880) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A, 139728972976896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'], 139728972979776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['first'].weight, 139732773125504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B, 139728972984528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B['first'], 139728972975888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.base_layer, 139777662337328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout, 139728972986352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['first'], 139728972988176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k, 139777662337232) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.weight, 139777660414656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q, 139777662337088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.weight, 139777664172096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.processor, 139777663631264) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm, 139777663630928) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm, 139777663631072) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu, 139777663630976) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear, 139728972338208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A, 139728972730464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'], 139728972729936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['first'].weight, 139732771183632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B, 139728972742080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B['first'], 139728972729216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.base_layer, 139777663631024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout, 139728972335136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['first'], 139728972337392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp, 139777663631168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp, 139728972727200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A, 139728972296112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'], 139728972286656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['first'].weight, 139732771176512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B, 139728972287280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B['first'], 139728972286944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.base_layer, 139777663631120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout, 139728972286176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['first'], 139728972287808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out, 139728972288768) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A, 139728972287760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['first'], 139728972285120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['first'].weight, 139732773120304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B, 139728972288816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B['first'], 139728972288960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.base_layer, 139777663631216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout, 139728972288384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout['first'], 139728972287376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3], 139777663630880) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn, 139777662337808) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k, 139728973077600) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A, 139728973136992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'], 139728973150336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['first'].weight, 139732772993392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B, 139728973145872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B['first'], 139728973152016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.base_layer, 139777662337952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout, 139728973080816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['first'], 139728973083552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q, 139728973080432) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A, 139728973086624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'], 139728973077360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['first'].weight, 139732772980192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B, 139728973082400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B['first'], 139728973071840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.base_layer, 139777662338048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout, 139728973079136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['first'], 139728973080912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v, 139728973140496) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A, 139728973149616) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'], 139728973147312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['first'].weight, 139732772992432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B, 139728973151872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B['first'], 139728973142512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.base_layer, 139777662338096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout, 139728973139872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['first'], 139728973148752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k, 139777662338000) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.weight, 139777660434480) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q, 139777662337856) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.weight, 139777664183776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.processor, 139777662337760) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm, 139777662337424) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm, 139777662337568) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu, 139777662337472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear, 139728972975984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A, 139728972979200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'], 139728973083696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['first'].weight, 139732773123504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B, 139728972985968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B['first'], 139728973080048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.base_layer, 139777662337520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout, 139728972983280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['first'], 139728972985344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp, 139777662337664) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp, 139728973081968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A, 139728973081488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'], 139728973075536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['first'].weight, 139732773117104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B, 139728973082160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B['first'], 139728973075152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.base_layer, 139777662337616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout, 139728973073184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['first'], 139728973082304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out, 139728973078032) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A, 139728973077648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['first'], 139728973080864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['first'].weight, 139732773112144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B, 139728973078800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B['first'], 139728973077312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.base_layer, 139777662337712) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout, 139728973078992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout['first'], 139728973077408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4], 139777662337376) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn, 139777662338576) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k, 139728972269024) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A, 139728972272432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'], 139728972273824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['first'].weight, 139732772787344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B, 139728972271184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B['first'], 139728972276032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.base_layer, 139777662338720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout, 139728972277232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['first'], 139728972275120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q, 139728974045168) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A, 139728974038208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'], 139728972268784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['first'].weight, 139732772791584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B, 139728974042096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B['first'], 139728972268544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.base_layer, 139777662338816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout, 139728974048432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['first'], 139728974048384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v, 139728972275936) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A, 139728972269072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'], 139728972277952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['first'].weight, 139732772795744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B, 139728972269648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B['first'], 139728972277664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.base_layer, 139777662338864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout, 139728972276752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['first'], 139728972277520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k, 139777662338768) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.weight, 139777664176976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q, 139777662338624) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.weight, 139777660379648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.processor, 139777662338528) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm, 139777662338192) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm, 139777662338336) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu, 139777662338240) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear, 139728973143520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A, 139728973140736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'], 139728973145488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['first'].weight, 139732772989232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B, 139728973144384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B['first'], 139728973145392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.base_layer, 139777662338288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout, 139728973145680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['first'], 139728973136224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp, 139777662338432) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp, 139728973147696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A, 139728973147024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'], 139728973150432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['first'].weight, 139732772982992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B, 139728973144144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B['first'], 139728973149568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.base_layer, 139777662338384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout, 139728973138240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['first'], 139728973139776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out, 139728973149664) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A, 139728973146016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['first'], 139728974050448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['first'].weight, 139732772790544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B, 139728973146832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B['first'], 139728974050736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.base_layer, 139777662338480) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout, 139728973152160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout['first'], 139728973149376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5], 139777662338144) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn, 139777662339344) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k, 139728972551920) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A, 139728972551152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'], 139728972558928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['first'].weight, 139732772628624) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B, 139728972553648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B['first'], 139728972549040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.base_layer, 139777662339488) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout, 139728972552592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['first'], 139728972553744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q, 139728973662624) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A, 139728972557104) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'], 139728972552112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['first'].weight, 139732772624144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B, 139728972553936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B['first'], 139728972552928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.base_layer, 139777662339584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout, 139728972556336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['first'], 139728972555232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v, 139728972554992) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A, 139728972816272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'], 139728972810416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['first'].weight, 139732772628704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B, 139728972813872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B['first'], 139728972813632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.base_layer, 139777662339632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout, 139728972558256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['first'], 139728972555664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k, 139777662339536) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.weight, 139781198839008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q, 139777662339392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.weight, 139777664579472) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.processor, 139777662339296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm, 139777662338960) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm, 139777662339104) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu, 139777662339008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear, 139728972282128) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A, 139728972268880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'], 139728972272912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['first'].weight, 139732772784384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B, 139728972270752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B['first'], 139728972270368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.base_layer, 139777662339056) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout, 139728972277568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['first'], 139728972275888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp, 139777662339200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp, 139728972271328) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A, 139728972269888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'], 139728973661712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['first'].weight, 139732772783744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B, 139728972276704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B['first'], 139728973672848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.base_layer, 139777662339152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout, 139728972272144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['first'], 139728972269168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out, 139728973673328) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A, 139728973661136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['first'], 139728973671984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['first'].weight, 139732772621744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B, 139728973674384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B['first'], 139728973662816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.base_layer, 139777662339248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout, 139728973676064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout['first'], 139728973668384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6], 139777662338912) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn, 139777662340112) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k, 139728971307072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A, 139728971302176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'], 139728971316144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['first'].weight, 139732772514816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B, 139728971307168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B['first'], 139728971316048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.base_layer, 139777662340256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout, 139728971310960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['first'], 139728971309616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q, 139728971315280) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A, 139728971314416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'], 139728971301888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['first'].weight, 139732772504016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B, 139728971308080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B['first'], 139728971307600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.base_layer, 139777662340352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout, 139728971312352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['first'], 139728971314848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v, 139728971204736) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A, 139728971213184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'], 139728971212128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['first'].weight, 139732772518656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B, 139728971212224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B['first'], 139728971208912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.base_layer, 139777662340400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout, 139728971209968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['first'], 139728971213664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k, 139777662340304) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.weight, 139777664186496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q, 139777662340160) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.weight, 139777660410336) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.processor, 139777662340064) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm, 139777662339728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm, 139777662339872) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu, 139777662339776) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear, 139728972812576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A, 139728972811568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'], 139728972814784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['first'].weight, 139732772624704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B, 139728972811376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B['first'], 139728972817328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.base_layer, 139777662339824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout, 139728972813008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['first'], 139728972813200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp, 139777662339968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp, 139728972818288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A, 139728972818000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'], 139728972627200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['first'].weight, 139732772628784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B, 139728972623360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B['first'], 139728972614432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.base_layer, 139777662339920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout, 139728972818192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['first'], 139728972817424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out, 139728972618224) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A, 139728971314128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['first'], 139728971315184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['first'].weight, 139732772621824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B, 139728971310624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B['first'], 139728971309904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.base_layer, 139777662340016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout, 139728971302416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout['first'], 139728972612752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7], 139777662339680) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn, 139777662340880) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k, 139728971383280) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A, 139728971281456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'], 139728971272336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['first'].weight, 139732772400128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B, 139728971273584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B['first'], 139728971272048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.base_layer, 139777662341024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout, 139728971387408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['first'], 139728971387936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q, 139728971398544) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A, 139728971396816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'], 139728971389952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['first'].weight, 139732772390608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B, 139728971397728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B['first'], 139728971383856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.base_layer, 139777662341120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout, 139728971397920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['first'], 139728971398592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v, 139728971281984) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A, 139728971282608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'], 139728972656128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['first'].weight, 139732772403328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B, 139728971283136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B['first'], 139728972651136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.base_layer, 139777662341168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout, 139728971279536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['first'], 139728971273680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k, 139777662341072) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.weight, 139777664181696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q, 139777662340928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.weight, 139777664579392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.processor, 139777662340832) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm, 139777662340496) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm, 139777662340640) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu, 139777662340544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear, 139728971250560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A, 139728971235728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'], 139728971389808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['first'].weight, 139732772508256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B, 139728971250752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B['first'], 139728971393504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.base_layer, 139777662340592) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout, 139728971239376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['first'], 139728971240816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp, 139777662340736) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp, 139728971392352) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A, 139728971393696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'], 139728971393744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['first'].weight, 139732772511296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B, 139728971393168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B['first'], 139728971398928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.base_layer, 139777662340688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout, 139728971392208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['first'], 139728971393312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out, 139728971391584) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A, 139728971390528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['first'], 139728971390480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['first'].weight, 139732772505376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B, 139728971397968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B['first'], 139728971399120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.base_layer, 139777662340784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout, 139728971391920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout['first'], 139728971392880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8], 139777662340448) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn, 139777662341648) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k, 139728972261088) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A, 139728972936480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'], 139728972937104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['first'].weight, 139732774271824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B, 139728972936048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B['first'], 139728972929232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.base_layer, 139777662341792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout, 139728972926688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['first'], 139728972259408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q, 139728972251296) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A, 139728972259696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'], 139728972265936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['first'].weight, 139732774272304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B, 139728972260464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B['first'], 139728972265264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.base_layer, 139777662341888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout, 139728972251440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['first'], 139728972251392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v, 139728972929712) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A, 139728972930288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'], 139728972835008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['first'].weight, 139732774272224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B, 139728972936864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B['first'], 139728972825264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.base_layer, 139777662341936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout, 139728972938352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['first'], 139728972929856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k, 139777662341840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.weight, 139777664477008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q, 139777662341696) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.weight, 139777664475008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.processor, 139777662341600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm, 139777662341264) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm, 139777662341408) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu, 139777662341312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear, 139728972659152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A, 139728972660640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'], 139728972658576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['first'].weight, 139732772401728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B, 139728972659584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B['first'], 139728972652000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.base_layer, 139777662341360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout, 139728972644464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['first'], 139728972652240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp, 139777662341504) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp, 139728972652096) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A, 139728972640800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'], 139728972641136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['first'].weight, 139732772393808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B, 139728972632928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B['first'], 139728972643920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.base_layer, 139777662341456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout, 139728972646048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['first'], 139728972659008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out, 139728972631008) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A, 139728972632160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['first'], 139728972254800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['first'].weight, 139732774265344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B, 139728972642336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B['first'], 139728972254464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.base_layer, 139777662341552) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout, 139728972643728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout['first'], 139728972630288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9], 139777662341216) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn, 139777662342416) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k, 139728971115264) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A, 139728971113536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'], 139728971106432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['first'].weight, 139732774135072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B, 139728971106624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B['first'], 139728971107152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.base_layer, 139777662342560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout, 139728971117088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['first'], 139728971115216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q, 139728971119152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A, 139728971115552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'], 139728971114928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['first'].weight, 139732774260384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B, 139728971109888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B['first'], 139728971116224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.base_layer, 139777662342656) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout, 139728971118672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['first'], 139728971118144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v, 139728971106096) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A, 139728971106528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'], 139728971105664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['first'].weight, 139732774137312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B, 139728971119248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B['first'], 139728971119056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.base_layer, 139777662342704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout, 139728971105856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['first'], 139728971107008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k, 139777662342608) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.weight, 139777664476128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q, 139777662342464) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.weight, 139777664479568) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.processor, 139777662342368) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm, 139777662342032) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm, 139777662342176) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu, 139777662342080) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear, 139728972826368) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A, 139728972833904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'], 139728971175472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['first'].weight, 139732774268304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B, 139728972829728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B['first'], 139728971175616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.base_layer, 139777662342128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout, 139728972836016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['first'], 139728972836928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp, 139777662342272) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp, 139728971175904) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A, 139728971181712) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'], 139728971111760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['first'].weight, 139732774270144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B, 139728971185792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B['first'], 139728971116944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.base_layer, 139777662342224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout, 139728971184592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['first'], 139728971170864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out, 139728971105280) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A, 139728971117856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['first'], 139728971115456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['first'].weight, 139732774263744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B, 139728971120064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B['first'], 139728971116032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.base_layer, 139777662342320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout, 139728971119104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout['first'], 139728971120256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10], 139777662341984) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn, 139777662343184) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k, 139728971470960) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A, 139728971468704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'], 139728971471344) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['first'].weight, 139732773997200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B, 139728971470816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B['first'], 139728971473120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.base_layer, 139777662343328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout, 139728971471008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['first'], 139728971469856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q, 139728971466352) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A, 139728971471488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'], 139728971466832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['first'].weight, 139732774129232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B, 139728971468944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B['first'], 139728971466784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.base_layer, 139777662343424) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout, 139728971466064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['first'], 139728971467264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v, 139728971472544) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A, 139728971473216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'], 139728971473408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['first'].weight, 139732773995200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B, 139728971473024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B['first'], 139728971474992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.base_layer, 139777662343472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout, 139728971472448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['first'], 139728971471728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k, 139777662343376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.weight, 139777659800128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q, 139777662343232) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.weight, 139777664329632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.processor, 139777662343136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm, 139777662342800) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm, 139777662342944) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu, 139777662342848) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear, 139728971114256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A, 139728971117040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'], 139728971115120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['first'].weight, 139732774128592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B, 139728971116176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B['first'], 139728971116128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.base_layer, 139777662342896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout, 139728971114736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['first'], 139728971114880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp, 139777662343040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp, 139728971106480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A, 139728971106288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'], 139728971469280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['first'].weight, 139732774140112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B, 139728971105808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B['first'], 139728971469232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.base_layer, 139777662342992) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout, 139728971106768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['first'], 139728971104800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out, 139728971469568) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A, 139728971469808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['first'], 139728971470336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['first'].weight, 139732774137392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B, 139728971470096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B['first'], 139728971470672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.base_layer, 139777662343088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout, 139728971469520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout['first'], 139728971468656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11], 139777662342752) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn, 139777662343952) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k, 139728972035072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A, 139728972026192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'], 139728972026384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['first'].weight, 139732773809552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B, 139728972030848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B['first'], 139728972031568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.base_layer, 139777662344096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout, 139728972031328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['first'], 139728972035408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q, 139728971480368) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A, 139728971480560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'], 139728972027104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['first'].weight, 139732773998800) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B, 139728971480704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B['first'], 139728972033392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.base_layer, 139777662344192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout, 139728971480272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['first'], 139728971480320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v, 139728972022112) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A, 139728972022304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'], 139728971167680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['first'].weight, 139732773806272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B, 139728971155296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B['first'], 139728971169024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.base_layer, 139777662344240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout, 139728972022016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['first'], 139728972022064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k, 139777662344144) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.weight, 139777660418096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q, 139777662344000) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.weight, 139777659854400) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.processor, 139777662343904) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm, 139777662343568) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm, 139777662343712) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu, 139777662343616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear, 139728971475952) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A, 139728971476624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'], 139728971476672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['first'].weight, 139732773995600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B, 139728971477008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B['first'], 139728971479264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.base_layer, 139777662343664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout, 139728971473888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['first'], 139728971473552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp, 139777662343808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp, 139728971479600) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A, 139728971478256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'], 139728971477824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['first'].weight, 139732774007120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B, 139728971479168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B['first'], 139728971477920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.base_layer, 139777662343760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout, 139728971479696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['first'], 139728971479648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out, 139728971464816) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A, 139728971478784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['first'], 139728971479936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['first'].weight, 139732774006560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B, 139728971478640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B['first'], 139728971479984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.base_layer, 139777662343856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout, 139728971477344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout['first'], 139728971477680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12], 139777662343520) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn, 139777662344720) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k, 139728976964640) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A, 139728976963872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'], 139728976969296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['first'].weight, 139732773670400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B, 139728976954656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B['first'], 139728976964928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.base_layer, 139777662344864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout, 139728976961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['first'], 139728976962768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q, 139728972691392) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A, 139728976524768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'], 139728976963824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['first'].weight, 139732773807312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B, 139728976517616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B['first'], 139728976962192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.base_layer, 139777662344960) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout, 139728976524384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['first'], 139728976522272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v, 139728976959264) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A, 139728976959168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'], 139728976958880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['first'].weight, 139732773670560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B, 139728976957776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B['first'], 139728976958256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.base_layer, 139777662345008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout, 139728976960512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['first'], 139728976958400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k, 139777662344912) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.weight, 139777664575232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q, 139777662344768) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.weight, 139781198835088) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.processor, 139777662344672) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm, 139777662344336) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm, 139777662344480) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu, 139777662344384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear, 139728971163312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A, 139728972861584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'], 139728972869360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['first'].weight, 139732773809072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B, 139728972869984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B['first'], 139728972863504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.base_layer, 139777662344432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout, 139728971169072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['first'], 139728971164224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp, 139777662344576) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp, 139728972862976) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A, 139728978267248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'], 139728978264752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['first'].weight, 139732773797952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B, 139728978270704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B['first'], 139728978275888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.base_layer, 139777662344528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout, 139728972863120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['first'], 139728972867536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out, 139728978280400) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A, 139728978277904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['first'], 139728978270848) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['first'].weight, 139732773804752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B, 139728978268976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B['first'], 139728978275456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.base_layer, 139777662344624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout, 139728978274736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout['first'], 139728978276032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13], 139777662344288) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn, 139777662345488) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k, 139728976676128) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A, 139728976687312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'], 139728977847504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['first'].weight, 139732773490736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B, 139728976682896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B['first'], 139728977849472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.base_layer, 139777662345632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout, 139728976675408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['first'], 139728976676080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q, 139728978261136) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A, 139728978256720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'], 139728976675840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['first'].weight, 139732773672640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B, 139728978254032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B['first'], 139728976676224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.base_layer, 139777662345728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout, 139728978258016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['first'], 139728978257728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v, 139728977839920) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A, 139728977850768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'], 139728978370160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['first'].weight, 139732773488736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B, 139728977849424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B['first'], 139728978365552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.base_layer, 139777662345776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout, 139728977850432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['first'], 139728977846448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k, 139777662345680) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.weight, 139777664581312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q, 139777662345536) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.weight, 139777659801808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.processor, 139777662345440) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm, 139777662345104) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm, 139777662345248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu, 139777662345152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear, 139728976958160) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A, 139728976958640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'], 139728976953504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['first'].weight, 139732773680560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B, 139728976956144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B['first'], 139728976965888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.base_layer, 139777662345200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout, 139728976957248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['first'], 139728976958496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp, 139777662345344) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp, 139728976968096) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A, 139728976967808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'], 139728976964160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['first'].weight, 139732773669040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B, 139728976968528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B['first'], 139728976965552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.base_layer, 139777662345296) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout, 139728976953600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['first'], 139728976966176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out, 139728976969152) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A, 139728976965744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['first'], 139728978262960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['first'].weight, 139732773677360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B, 139728978254800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B['first'], 139728978255856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.base_layer, 139777662345392) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout, 139728976954416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout['first'], 139728976955136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14], 139777662345056) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn, 139777662346256) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k, 139728976866912) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A, 139728976857312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'], 139728976857888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['first'].weight, 139732773349520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B, 139728976861920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B['first'], 139728976857984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.base_layer, 139777662346400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout, 139728976866528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['first'], 139728976866768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q, 139728977057712) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A, 139728977062128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'], 139728976869312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['first'].weight, 139732773348000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B, 139728977064912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B['first'], 139728976855728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.base_layer, 139777662346496) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout, 139728977052288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['first'], 139728977053104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v, 139728976856304) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A, 139728976856928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'], 139728976867008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['first'].weight, 139732773352480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B, 139728976855824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B['first'], 139728976865376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.base_layer, 139777662346544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout, 139728976858080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['first'], 139728976856400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k, 139777662346448) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.weight, 139777664751776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q, 139777662346304) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.weight, 139777664573952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.processor, 139777662346208) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm, 139777662345872) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm, 139777662346016) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu, 139777662345920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear, 139728978363056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A, 139728978365984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'], 139728978375584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['first'].weight, 139732773486656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B, 139728978376064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B['first'], 139728978373904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.base_layer, 139777662345968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout, 139728978373424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['first'], 139728978375680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp, 139777662346112) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp, 139728978369248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A, 139728978363008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'], 139728978366896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['first'].weight, 139732773500336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B, 139728978371024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B['first'], 139728978377216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.base_layer, 139777662346064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout, 139728978373568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['first'], 139728978369584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out, 139728978271616) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A, 139728977064672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['first'], 139728977055312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['first'].weight, 139732773495056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B, 139728977054832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B['first'], 139728977060976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.base_layer, 139777662346160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout, 139728977052864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout['first'], 139728977064384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15], 139777662345824) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn, 139777662347024) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k, 139728976897328) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A, 139728976891760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'], 139728976902320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['first'].weight, 139732775303376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B, 139728976895024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B['first'], 139728976902224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.base_layer, 139777662347168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout, 139728976897280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['first'], 139728976897184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q, 139728976717392) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A, 139728976716528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'], 139728976896032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['first'].weight, 139732773344240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B, 139728976717584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B['first'], 139728976894928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.base_layer, 139777662347264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout, 139728976719936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['first'], 139728976714368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v, 139728978363584) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A, 139728977147120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'], 139728977139152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['first'].weight, 139732775295136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B, 139728977145968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B['first'], 139728977139536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.base_layer, 139777662347312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout, 139728977147264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['first'], 139728977143904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k, 139777662347216) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.weight, 139777659855920) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q, 139777662347072) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.weight, 139777660118384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.processor, 139777662346976) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm, 139777662346640) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm, 139777662346784) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu, 139777662346688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear, 139728976862928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A, 139728976867056) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'], 139728976856688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['first'].weight, 139732773343760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B, 139728976862016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B['first'], 139728976857360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.base_layer, 139777662346736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout, 139728976866816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['first'], 139728976863024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp, 139777662346880) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp, 139728976866864) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A, 139728976855440) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'], 139728976772160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['first'].weight, 139732773354880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B, 139728976759488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B['first'], 139728976765440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.base_layer, 139777662346832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout, 139728976856256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['first'], 139728976861632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out, 139728976761504) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A, 139728976765152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['first'], 139728976758048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['first'].weight, 139732773339200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B, 139728976760208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B['first'], 139728976720176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.base_layer, 139777662346928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout, 139728976772832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout['first'], 139728976761216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16], 139777662346592) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn, 139777662347792) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k, 139728977139200) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A, 139728977138624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'], 139728977086208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['first'].weight, 139732775176848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B, 139728977146208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B['first'], 139728977087120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.base_layer, 139777662347936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout, 139728977140352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['first'], 139728977136512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q, 139728977143232) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A, 139728977145056) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'], 139728977145296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['first'].weight, 139732775295456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B, 139728977142320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B['first'], 139728977142416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.base_layer, 139777662348032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout, 139728977140880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['first'], 139728977140688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v, 139728977090576) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A, 139728977091632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'], 139728977088656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['first'].weight, 139732775184128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B, 139728977097056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B['first'], 139728977090144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.base_layer, 139777662348080) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout, 139728977092496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['first'], 139728977090432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k, 139777662347984) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.weight, 139777664323712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q, 139777662347840) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.weight, 139777664746576) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.processor, 139777662347744) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm, 139777662347408) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm, 139777662347552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu, 139777662347456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear, 139728977148032) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A, 139728977147936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'], 139728977148752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['first'].weight, 139732775299696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B, 139728977140784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B['first'], 139728977148800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.base_layer, 139777662347504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout, 139728977139248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['first'], 139728977148272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp, 139777662347648) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp, 139728977146544) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A, 139728977144624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'], 139728977141936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['first'].weight, 139732775302896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B, 139728977146064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B['first'], 139728977144096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.base_layer, 139777662347600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout, 139728977148848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['first'], 139728977145392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out, 139728977141264) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A, 139728977139680) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['first'], 139728977142560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['first'].weight, 139732775303616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B, 139728977139488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B['first'], 139728977141504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.base_layer, 139777662347696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout, 139728977137904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout['first'], 139728977140592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17], 139777662347360) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn, 139777662348560) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k, 139728976902608) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A, 139728984827088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'], 139728984950576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['first'].weight, 139732774993760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B, 139728984821040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B['first'], 139728984956576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.base_layer, 139777662348704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout, 139728984821952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['first'], 139728984822768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q, 139728977241440) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A, 139728977241584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'], 139728977238032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['first'].weight, 139732775179968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B, 139728977247056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B['first'], 139728977240192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.base_layer, 139777662348800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout, 139728977239856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['first'], 139728977241632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v, 139728984951008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A, 139728984950480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'], 139728984950096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['first'].weight, 139732774984080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B, 139728984961616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B['first'], 139728984165728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.base_layer, 139777662348848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout, 139728984962528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['first'], 139728984955616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k, 139777662348752) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.weight, 139777660124544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q, 139777662348608) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.weight, 139777659810208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.processor, 139777662348512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm, 139777662348176) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm, 139777662348320) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu, 139777662348224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear, 139728977091920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A, 139728977089664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'], 139728977084480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['first'].weight, 139732775189168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B, 139728977089856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B['first'], 139728977090768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.base_layer, 139777662348272) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout, 139728977086112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['first'], 139728977094896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp, 139777662348416) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp, 139728977092400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A, 139728977089136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'], 139728977238896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['first'].weight, 139732775176768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B, 139728977093696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B['first'], 139728977239232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.base_layer, 139777662348368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout, 139728977096960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['first'], 139728977093264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out, 139728977238848) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A, 139728977241008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['first'], 139728977242016) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['first'].weight, 139732775187248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B, 139728977238272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B['first'], 139728977241872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.base_layer, 139777662348464) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout, 139728977238944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout['first'], 139728977237456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18], 139777662348128) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn, 139777662349328) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k, 139728977232752) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A, 139728985170096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'], 139728985166784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['first'].weight, 139732774978400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B, 139728985172784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B['first'], 139728985167984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.base_layer, 139777662349472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout, 139728985057808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['first'], 139728985053776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q, 139728985001600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A, 139728985003952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'], 139728985004720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['first'].weight, 139732774984000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B, 139728985013456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B['first'], 139728985004288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.base_layer, 139777662349568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout, 139728985011248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['first'], 139728985013024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v, 139728985237456) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A, 139728985237648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'], 139728985267824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['first'].weight, 139732774872592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B, 139728985228096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B['first'], 139728985269504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.base_layer, 139777662349616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout, 139728985239040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['first'], 139728985239616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k, 139777662349520) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.weight, 139777660373728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q, 139777662349376) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.weight, 139777660372208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.processor, 139777662349280) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm, 139777662348944) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm, 139777662349088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu, 139777662348992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear, 139728984176528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A, 139728984167408) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'], 139728985418160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['first'].weight, 139732774984240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B, 139728984164576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B['first'], 139728985418208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.base_layer, 139777662349040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout, 139728984170336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['first'], 139728984168368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp, 139777662349184) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp, 139728985421472) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A, 139728985413264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'], 139728985407840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['first'].weight, 139732774990240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B, 139728985412448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B['first'], 139728985408272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.base_layer, 139777662349136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout, 139728985420512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['first'], 139728985412208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out, 139728985413552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A, 139728985412016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['first'], 139728985002752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['first'].weight, 139732774989520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B, 139728985421232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B['first'], 139728984998432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.base_layer, 139777662349232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout, 139728985409040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout['first'], 139728985418880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19], accessed_by=GetItemGuardAccessor(19) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19], 139777662348896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn, 139777662350096) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k, 139728983778512) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A, 139728983777168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'], 139728985581184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['first'].weight, 139732774679904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B, 139728983772560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B['first'], 139728985582720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.base_layer, 139777662350240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout, 139728983776688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['first'], 139728983780720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q, 139728985529056) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A, 139728985538224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'], 139728983773760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['first'].weight, 139732774864272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B, 139728983785088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B['first'], 139728983771600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.base_layer, 139777662350336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout, 139728985534384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['first'], 139728985527904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v, 139728985587472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A, 139728985577920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'], 139728984126416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['first'].weight, 139732774666384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B, 139728985571776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B['first'], 139728984117056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.base_layer, 139777662350384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout, 139728985584640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['first'], 139728985585024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k, 139777662350288) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.weight, 139777664712304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q, 139777662350144) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.weight, 139777660377488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.processor, 139777662350048) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm, 139777662349712) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm, 139777662349856) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu, 139777662349760) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear, 139728985268016) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A, 139728985267248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'], 139728985378096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['first'].weight, 139732774869232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B, 139728985271568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B['first'], 139728985378240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.base_layer, 139777662349808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout, 139728985264512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['first'], 139728985264176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp, 139777662349952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp, 139728985378528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A, 139728985482976) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'], 139728985473712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['first'].weight, 139732774863232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B, 139728985483936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B['first'], 139728985480768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.base_layer, 139777662349904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout, 139728985390288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['first'], 139728985384096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out, 139728985473616) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A, 139728985479184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['first'], 139728985535200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['first'].weight, 139732774867312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B, 139728985531264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B['first'], 139728985528192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.base_layer, 139777662350000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout, 139728985482832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout['first'], 139728985484560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20], accessed_by=GetItemGuardAccessor(20) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20], 139777662349664) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn, 139777662350864) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k, 139728985675840) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A, 139728985674784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'], 139729027944448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['first'].weight, 139732774530928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B, 139728985681696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B['first'], 139729027954864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.base_layer, 139777662351008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout, 139728985676128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['first'], 139728985677760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q, 139728985775152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A, 139728985772128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'], 139728985684240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['first'].weight, 139732774529648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B, 139728985783840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B['first'], 139728985671568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.base_layer, 139777662351104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout, 139728985776448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['first'], 139728985775488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v, 139728985014224) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A, 139729027839504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'], 139729027738944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['first'].weight, 139732774522208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B, 139729027836960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B['first'], 139729027732896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.base_layer, 139777662351152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout, 139729027839168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['first'], 139729027834800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k, 139777662351056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.weight, 139777664182496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q, 139777662350912) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.weight, 139777664729392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.processor, 139777662350816) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm, 139777662350480) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm, 139777662350624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu, 139777662350528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear, 139728984115040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A, 139728984117104) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'], 139728984118736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['first'].weight, 139732774676304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B, 139728984115184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B['first'], 139728984119264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.base_layer, 139777662350576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout, 139728984118496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['first'], 139728984113888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp, 139777662350720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp, 139728984121712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A, 139728984121424) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'], 139728984122096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['first'].weight, 139732774677104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B, 139728984127856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B['first'], 139728984115664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.base_layer, 139777662350672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout, 139728984123440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['first'], 139728984128480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out, 139728985725904) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A, 139728985726672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['first'], 139728985768144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['first'].weight, 139732774670464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B, 139728985730080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B['first'], 139728985769632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.base_layer, 139777662350768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout, 139728985725760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout['first'], 139728985719520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21], accessed_by=GetItemGuardAccessor(21) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21], 139777662350432) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn, 139777662351632) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k, 139729028935408) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A, 139729071121568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'], 139729070985024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['first'].weight, 139732774372912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B, 139729071115664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B['first'], 139729070826272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.base_layer, 139777662351776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout, 139729071261584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['first'], 139729071275936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q, 139729027941952) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A, 139729070046176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'], 139729071467120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['first'].weight, 139732774380352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B, 139729071757088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B['first'], 139729071457136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.base_layer, 139777662351872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout, 139729070400864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['first'], 139729070495760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v, 139729071458816) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A, 139729073486720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'], 139729072471008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['first'].weight, 139732774385472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B, 139729073485520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B['first'], 139729072464864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.base_layer, 139777662351920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout, 139729072770544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['first'], 139729072767280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k, 139777662351824) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.weight, 139777660377808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q, 139777662351680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.weight, 139777664725328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.processor, 139777662351584) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm, 139777662351248) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm, 139777662351392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu, 139777662351296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear, 139729031521440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A, 139729030869728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'], 139729029444896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['first'].weight, 139732774533328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B, 139729073071840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B['first'], 139729029445472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.base_layer, 139777662351344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout, 139729030517024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['first'], 139729030513088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp, 139777662351488) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp, 139729029447152) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A, 139729029435152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'], 139729029008784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['first'].weight, 139732774533728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B, 139729029300480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B['first'], 139729029016944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.base_layer, 139777662351440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout, 139729029446864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['first'], 139729029446960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out, 139729029021120) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A, 139729029123808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['first'], 139729028933680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['first'].weight, 139732774527488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B, 139729029121936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B['first'], 139729028939584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.base_layer, 139777662351536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout, 139729029132256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout['first'], 139729029020208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22], accessed_by=GetItemGuardAccessor(22) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22], 139777662351200) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn, 139777662352400) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k, 139729075138144) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A, 139729075128688) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'], 139729076783424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['first'].weight, 139732776313424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B, 139729076787024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B['first'], 139729076782992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.base_layer, 139777662352544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout, 139729075138480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['first'], 139729075138672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q, 139729073604208) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A, 139729075662192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'], 139729075333744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['first'].weight, 139732774373232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B, 139729075661328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B['first'], 139729075139728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.base_layer, 139777662352640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout, 139729075660608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['first'], 139729075660224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v, 139729078054256) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A, 139729078057568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'], 139728986596512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['first'].weight, 139732776314464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B, 139729078050320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B['first'], 139728986600496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.base_layer, 139777662352688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout, 139729078055120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['first'], 139729078054064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k, 139777662352592) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.weight, 139777664757856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q, 139777662352448) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.weight, 139777660378528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.processor, 139777662352352) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm, 139777662352016) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm, 139777662352160) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu, 139777662352064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear, 139729072306688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A, 139729072306448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'], 139729072041376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['first'].weight, 139732774373392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B, 139729072306208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B['first'], 139729072042480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.base_layer, 139777662352112) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout, 139729072294208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['first'], 139729072299872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp, 139777662352256) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp, 139729072038160) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A, 139729071776400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'], 139729073605024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['first'].weight, 139732774387472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B, 139729071770016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B['first'], 139729073604496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.base_layer, 139777662352208) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout, 139729071783648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['first'], 139729071783456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out, 139729070823680) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A, 139729073304432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['first'], 139729073161408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['first'].weight, 139732774378352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B, 139729073304528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B['first'], 139729073166112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.base_layer, 139777662352304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout, 139729073299296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout['first'], 139729073293056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23], accessed_by=GetItemGuardAccessor(23) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23], 139777662351968) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn, 139777662353168) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k, 139729076457856) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A, 139729076456224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'], 139729076368336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['first'].weight, 139732776185312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B, 139729076457232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B['first'], 139729076362528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.base_layer, 139777661026368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout, 139729076466736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['first'], 139729076454496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q, 139729074952976) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A, 139729074950432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'], 139729076200800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['first'].weight, 139732776182752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B, 139729074946496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B['first'], 139729076469040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.base_layer, 139777662353312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout, 139729074961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['first'], 139729074959648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v, 139729076356960) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A, 139729077922128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'], 139729077812144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['first'].weight, 139732776186032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B, 139729077927504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B['first'], 139729077812912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.base_layer, 139777661026416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout, 139729077912144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['first'], 139729076358640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k, 139777662353360) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.weight, 139777660445824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q, 139777662353216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.weight, 139777664332992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.processor, 139777662353120) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm, 139777662352784) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm, 139777662352928) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu, 139777662352832) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear, 139728986592768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A, 139728986129744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'], 139728986138480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['first'].weight, 139732776319184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B, 139728986142176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B['first'], 139728986141792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.base_layer, 139777662352880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout, 139728986590560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['first'], 139728986591184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp, 139777662353024) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp, 139728986134400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A, 139728986136272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'], 139728986503632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['first'].weight, 139732776312464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B, 139728986134736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B['first'], 139728986504400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.base_layer, 139777662352976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout, 139728986133104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['first'], 139728986133200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out, 139728986503344) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A, 139728986498544) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['first'], 139729076923616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['first'].weight, 139732776175312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B, 139729076928080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B['first'], 139729074955616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.base_layer, 139777662353072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout, 139728986495088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout['first'], 139728986499648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24], accessed_by=GetItemGuardAccessor(24) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24], 139777662352736) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn, 139777661026896) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k, 139729080562160) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A, 139729080563120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'], 139729080553424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['first'].weight, 139732776045280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B, 139729080563600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B['first'], 139729080560960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.base_layer, 139777661027040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout, 139729080562592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['first'], 139729080562640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q, 139729079640528) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A, 139729079641824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'], 139729080561488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['first'].weight, 139732776057600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B, 139729079632080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B['first'], 139729080564896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.base_layer, 139777661027136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout, 139729079637408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['first'], 139729079637936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v, 139729080553184) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A, 139729080557936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'], 139729080555488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['first'].weight, 139732776052720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B, 139729080552992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B['first'], 139729080558704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.base_layer, 139777661027184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout, 139729080562496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['first'], 139729080558272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k, 139777661027088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.weight, 139777660322976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q, 139777661026944) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.weight, 139777664578432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.processor, 139777661026848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm, 139777661026512) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm, 139777661026656) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu, 139777661026560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear, 139729077811472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A, 139729077580800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'], 139729079643456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['first'].weight, 139732776187312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B, 139729077576864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B['first'], 139729079646096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.base_layer, 139777661026608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout, 139729077812960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['first'], 139729077811376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp, 139777661026752) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp, 139729079646912) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A, 139729079643840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'], 139729079633376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['first'].weight, 139732776181152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B, 139729079638608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B['first'], 139729079637456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.base_layer, 139777661026704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout, 139729079647344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['first'], 139729079647152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out, 139729079643936) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A, 139729079634960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['first'], 139729079639376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['first'].weight, 139732776177232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B, 139729079635728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B['first'], 139729079639568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.base_layer, 139777661026800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout, 139729079646384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout['first'], 139729079638560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25], accessed_by=GetItemGuardAccessor(25) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25], 139777661026464) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn, 139777661027664) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k, 139729081832768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A, 139729081831280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'], 139729082851552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['first'].weight, 139732775842048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B, 139729081840016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B['first'], 139729082858752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.base_layer, 139777661027808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout, 139729081828112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['first'], 139729081832288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q, 139729081832816) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A, 139729081828352) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'], 139729081829264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['first'].weight, 139732775842368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B, 139729081828640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B['first'], 139729081830368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.base_layer, 139777661027904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout, 139729081835792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['first'], 139729081835456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v, 139729082854144) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A, 139729082857168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'], 139729082851840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['first'].weight, 139732775844928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B, 139729082857936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B['first'], 139729082855632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.base_layer, 139777661027952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout, 139729082855536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['first'], 139729082854096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k, 139777661027856) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.weight, 139777664182896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q, 139777661027712) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.weight, 139777664566032) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.processor, 139777661027616) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm, 139777661027280) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm, 139777661027424) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu, 139777661027328) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear, 139729080550784) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A, 139729080552368) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'], 139729080552800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['first'].weight, 139732776056640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B, 139729080550736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B['first'], 139729080555680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.base_layer, 139777661027376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout, 139729080549728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['first'], 139729080550592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp, 139777661027520) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp, 139729080556496) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A, 139729081837952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'], 139729081839968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['first'].weight, 139732776048080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B, 139729081834208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B['first'], 139729081842224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.base_layer, 139777661027472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout, 139729080553328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['first'], 139729080555056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out, 139729081843424) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A, 139729081839584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['first'], 139729081836224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['first'].weight, 139732775839648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B, 139729081837040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B['first'], 139729081836368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.base_layer, 139777661027568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout, 139729081840832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout['first'], 139729081842800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26], accessed_by=GetItemGuardAccessor(26) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26], 139777661027232) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn, 139777661028432) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k, 139729085484240) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A, 139729085485200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'], 139729085488800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['first'].weight, 139732775716080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B, 139729085485680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B['first'], 139729085486112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.base_layer, 139777661028576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout, 139729085484672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['first'], 139729085484720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q, 139729085484576) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A, 139729085491008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'], 139729085483136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['first'].weight, 139732775720880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B, 139729085491488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B['first'], 139729085487456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.base_layer, 139777661028672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout, 139729085493600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['first'], 139729085491728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v, 139729085490720) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A, 139729085481072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'], 139729086478896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['first'].weight, 139732775726720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B, 139729085490288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B['first'], 139729086477984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.base_layer, 139777661028720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout, 139729085490624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['first'], 139729085492160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k, 139777661028624) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.weight, 139777661092384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q, 139777661028480) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.weight, 139777661093744) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.processor, 139777661028384) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm, 139777661028048) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm, 139777661028192) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu, 139777661028096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear, 139729082856928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A, 139729082852752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'], 139729082849536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['first'].weight, 139732775834928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B, 139729082850496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B['first'], 139729082849296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.base_layer, 139777661028144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout, 139729082857264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['first'], 139729082856832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp, 139777661028288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp, 139729082848528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A, 139729082843872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'], 139729082844400) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['first'].weight, 139732775728160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B, 139729082844112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B['first'], 139729082844544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.base_layer, 139777661028240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout, 139729082850304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['first'], 139729082848480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out, 139729082848720) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A, 139729082857840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['first'], 139729085497104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['first'].weight, 139732775716160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B, 139729085496000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B['first'], 139729085497152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.base_layer, 139777661028336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout, 139729082855920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout['first'], 139729082849152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27], accessed_by=GetItemGuardAccessor(27) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27], 139777661028000) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn, 139777661029200) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k, 139729087700448) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A, 139729087699008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'], 139729087700352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['first'].weight, 139732775595568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B, 139729087699296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B['first'], 139729087699680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.base_layer, 139777661029344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout, 139729087704720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['first'], 139729087704816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q, 139729073171488) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A, 139729087694208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'], 139729087708320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['first'].weight, 139732775590928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B, 139729087698864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B['first'], 139729087708656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.base_layer, 139777661029440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout, 139729087703568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['first'], 139729087707360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v, 139729087693392) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A, 139729087693536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'], 139729087696944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['first'].weight, 139732775597408) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B, 139729087693296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B['first'], 139729087697280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.base_layer, 139777661029488) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout, 139729087693008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['first'], 139729087692960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k, 139777661029392) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.weight, 139777661093504) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q, 139777661029248) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.weight, 139777661094864) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.processor, 139777661029152) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm, 139777661028816) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm, 139777661028960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu, 139777661028864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear, 139729086476304) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A, 139729086478752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'], 139729086480048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['first'].weight, 139732775721360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B, 139729086478608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B['first'], 139729086480144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.base_layer, 139777661028912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout, 139729086477696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['first'], 139729086477888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp, 139777661029056) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp, 139729086469584) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A, 139729086471840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'], 139729086464592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['first'].weight, 139732775716640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B, 139729086466560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B['first'], 139729086464544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.base_layer, 139777661029008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout, 139729086472224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['first'], 139729086472704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out, 139729086466416) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A, 139729086468816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['first'], 139729086471168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['first'].weight, 139732775595328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B, 139729086469248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B['first'], 139729086470400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.base_layer, 139777661029104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout, 139729086464112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout['first'], 139729086467136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28], accessed_by=GetItemGuardAccessor(28) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28], 139777661028768) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn, 139777661030016) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k, 139729091443536) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A, 139729092148384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'], 139729092139840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['first'].weight, 139732775433008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B, 139729092139648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B['first'], 139729092138400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.base_layer, 139777661030160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout, 139729092142528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['first'], 139729091438256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q, 139729091429616) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A, 139729091429280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'], 139729091433072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['first'].weight, 139732775432208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B, 139729091431680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B['first'], 139729091440272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.base_layer, 139777661030256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout, 139729091429712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['first'], 139729091428656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v, 139729092148624) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A, 139729092140992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'], 139729092144064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['first'].weight, 139732775432048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B, 139729092137152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B['first'], 139729092144160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.base_layer, 139777661030304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout, 139729092143968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['first'], 139729092148336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k, 139777661030208) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.weight, 139777661094624) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q, 139777661030064) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.weight, 139777661096064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.processor, 139777661029968) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm, 139777661029584) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm, 139777661029728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu, 139777661029632) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear, 139729087705872) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A, 139729087707792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'], 139729091436864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['first'].weight, 139732775584688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B, 139729087708272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B['first'], 139729091431440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.base_layer, 139777661029680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout, 139729087695792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['first'], 139729087700016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp, 139777661029872) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp, 139729091442720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A, 139729091439984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'], 139729091443872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['first'].weight, 139732775433568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B, 139729091439888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B['first'], 139729091439552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.base_layer, 139777661029776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout, 139729091442768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['first'], 139729091437872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out, 139729091434944) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A, 139729091435040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['first'], 139729091434896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['first'].weight, 139732775433088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B, 139729091434368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B['first'], 139729091433696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.base_layer, 139777661029920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout, 139729091437056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout['first'], 139729091436000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29], accessed_by=GetItemGuardAccessor(29) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29], 139777661029536) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn, 139777661030784) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k, 139729093490768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A, 139729093490144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'], 139729093478816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['first'].weight, 139732777406752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B, 139729093486544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B['first'], 139729093478288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.base_layer, 139777661030928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout, 139729093483760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['first'], 139729093484192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q, 139729093484624) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A, 139729093491872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'], 139729093486352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['first'].weight, 139732777406272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B, 139729093484768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B['first'], 139729093485776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.base_layer, 139777661031024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout, 139729093482800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['first'], 139729093489520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v, 139729093483472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A, 139729093484576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'], 139729093488608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['first'].weight, 139732777409072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B, 139729093481744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B['first'], 139729093484960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.base_layer, 139777661031072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout, 139729093481984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['first'], 139729093483712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k, 139777661030976) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.weight, 139777661095824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q, 139777661030832) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.weight, 139777661097184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.processor, 139777661030736) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm, 139777661030400) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm, 139777661030544) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu, 139777661030448) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear, 139729092135568) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A, 139729092139744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'], 139729092145744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['first'].weight, 139732775424048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B, 139729092141376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B['first'], 139729092146224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.base_layer, 139777661030496) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout, 139729092133264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['first'], 139729092137296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp, 139777661030640) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp, 139729092147040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A, 139729092147280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'], 139729092139072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['first'].weight, 139732777414032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B, 139729092137824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B['first'], 139729092139312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.base_layer, 139777661030592) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout, 139729092141808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['first'], 139729092147952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out, 139729092133552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A, 139729092134800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['first'], 139729093485392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['first'].weight, 139732777415952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B, 139729092132928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B['first'], 139729093491200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.base_layer, 139777661030688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout, 139729092133648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout['first'], 139729092133360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30], accessed_by=GetItemGuardAccessor(30) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30], 139777661030352) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn, 139777661031552) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k, 139729091698624) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A, 139729091700928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'], 139729093050176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['first'].weight, 139732777230048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B, 139729091692672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B['first'], 139729093047872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.base_layer, 139777661031696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout, 139729091697376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['first'], 139729091703616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q, 139729091695024) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A, 139729091697232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'], 139729091690896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['first'].weight, 139732777231568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B, 139729091696752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B['first'], 139729091690608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.base_layer, 139777661031792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout, 139729091695120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['first'], 139729091696464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v, 139729093047056) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A, 139729093043168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'], 139729093045232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['first'].weight, 139732777234288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B, 139729093043072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B['first'], 139729093045808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.base_layer, 139777661031840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout, 139729093047680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['first'], 139729093046960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k, 139777661031744) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.weight, 139777661096944) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q, 139777661031600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.weight, 139777661098304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.processor, 139777661031504) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm, 139777661031168) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm, 139777661031312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu, 139777661031216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear, 139729093480016) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A, 139729093480976) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'], 139729091706064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['first'].weight, 139732777415392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B, 139729093481024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B['first'], 139729091703664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.base_layer, 139777661031264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout, 139729093480064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['first'], 139729093479824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp, 139777661031408) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp, 139729091705296) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A, 139729091703232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'], 139729091694064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['first'].weight, 139732777407632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B, 139729091699968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B['first'], 139729091702320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.base_layer, 139777661031360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout, 139729091706736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['first'], 139729091705200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out, 139729091701456) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A, 139729091691472) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['first'], 139729091695312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['first'].weight, 139732777233248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B, 139729091690848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B['first'], 139729091693536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.base_layer, 139777661031456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout, 139729091700592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout['first'], 139729091696848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31], accessed_by=GetItemGuardAccessor(31) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31], 139777661031120) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn, 139777661032320) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k, 139729093175344) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A, 139729093165984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'], 139729093169584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['first'].weight, 139732777126384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B, 139729093170736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B['first'], 139729093167328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.base_layer, 139777661032464) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout, 139729093166224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['first'], 139729093176112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q, 139729093177696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A, 139729093178656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'], 139729093172848) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['first'].weight, 139732777127424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B, 139729093179136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B['first'], 139729093175536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.base_layer, 139777661032560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout, 139729093178128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['first'], 139729093178176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v, 139729093173328) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A, 139729093165696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'], 139729093174528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['first'].weight, 139732777133424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B, 139729093170448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B['first'], 139729093173952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.base_layer, 139777661032608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout, 139729093172944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['first'], 139729093171552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k, 139777661032512) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.weight, 139777661098064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q, 139777661032368) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.weight, 139777661099424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.processor, 139777661032272) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm, 139777661031936) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm, 139777661032080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu, 139777661031984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear, 139729093035440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A, 139729093035200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'], 139729093038896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['first'].weight, 139732777235568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B, 139729093034480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B['first'], 139729093037216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.base_layer, 139777661032032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout, 139729093043504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['first'], 139729093036352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp, 139777661032176) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp, 139729093040624) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A, 139729093037072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'], 139729093043456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['first'].weight, 139732777237008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B, 139729093040816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B['first'], 139729093047920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.base_layer, 139777661032128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout, 139729093039328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['first'], 139729093035872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out, 139729093035920) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A, 139729093172080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['first'], 139729093166656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['first'].weight, 139732777222848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B, 139729093168672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B['first'], 139729093175776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.base_layer, 139777661032224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout, 139729093046768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout['first'], 139729093042544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32], accessed_by=GetItemGuardAccessor(32) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32], 139777661031888) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn, 139777661033088) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k, 139729094389680) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A, 139729094390640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'], 139729094382960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['first'].weight, 139732776935856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B, 139729094393568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B['first'], 139729094379984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.base_layer, 139777661033232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout, 139729094391600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['first'], 139729094389104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q, 139729093649472) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A, 139729093650672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'], 139729094392272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['first'].weight, 139732776934976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B, 139729093641648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B['first'], 139729094390832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.base_layer, 139777661033328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout, 139729093649904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['first'], 139729093650240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v, 139729094381376) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A, 139729094387328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'], 139729094377584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['first'].weight, 139732776930256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B, 139729094383008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B['first'], 139729094378304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.base_layer, 139777661033376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout, 139729094381808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['first'], 139729094381520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k, 139777661033280) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.weight, 139777661099184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q, 139777661033136) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.weight, 139777661100544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.processor, 139777661033040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm, 139777661032704) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm, 139777661032848) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu, 139777661032752) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear, 139729093655856) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A, 139729093652448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'], 139729093653600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['first'].weight, 139732777135024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B, 139729093653264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B['first'], 139729093652400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.base_layer, 139777661032800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout, 139729093655808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['first'], 139729093644960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp, 139777661032944) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp, 139729093653168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A, 139729093647312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'], 139729093644000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['first'].weight, 139732777130544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B, 139729093654032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B['first'], 139729093643520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.base_layer, 139777661032896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout, 139729093653024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['first'], 139729093651056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out, 139729093644240) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A, 139729093644912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['first'], 139729093647984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['first'].weight, 139732776940656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B, 139729093647552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B['first'], 139729093643376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.base_layer, 139777661032992) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout, 139729093644624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout['first'], 139729093641552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33], accessed_by=GetItemGuardAccessor(33) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33], 139777661032656) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn, 139777661033856) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k, 139729092741776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A, 139729092749456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'], 139729092741968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['first'].weight, 139732776728944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B, 139729092744224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B['first'], 139729092741824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.base_layer, 139777661034000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout, 139729092742736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['first'], 139729092741728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q, 139729092747008) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A, 139729092747392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'], 139729092744464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['first'].weight, 139732776724864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B, 139729092745280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B['first'], 139729092744608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.base_layer, 139777661034096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout, 139729092747536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['first'], 139729092748112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v, 139729092749264) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A, 139729094561552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'], 139729094569328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['first'].weight, 139732776726544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B, 139729094560544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B['first'], 139729094568848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.base_layer, 139777661034144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout, 139729094571344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['first'], 139729094571824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k, 139777661034048) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.weight, 139777661100304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q, 139777661033904) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.weight, 139777661101664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.processor, 139777661033808) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm, 139777661033472) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm, 139777661033616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu, 139777661033520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear, 139729094379024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A, 139729094381232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'], 139729094392224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['first'].weight, 139732776939696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B, 139729094380560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B['first'], 139729094392176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.base_layer, 139777661033568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout, 139729094379120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['first'], 139729094380464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp, 139777661033712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp, 139729094393808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A, 139729092755120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'], 139729092749072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['first'].weight, 139732776932336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B, 139729092755168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B['first'], 139729092748688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.base_layer, 139777661033664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout, 139729092750752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['first'], 139729094390256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out, 139729092754064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A, 139729092742880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['first'], 139729092751568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['first'].weight, 139732776927696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B, 139729092749792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B['first'], 139729092740912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.base_layer, 139777661033760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout, 139729092753440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout['first'], 139729092753536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34], accessed_by=GetItemGuardAccessor(34) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34], 139777661033424) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn, 139777661034624) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k, 139729095747776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A, 139729095745136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'], 139729095743936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['first'].weight, 139732776524736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B, 139729095742592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B['first'], 139729095745424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.base_layer, 139777661034768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout, 139729095747488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['first'], 139729095747584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q, 139729095749120) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A, 139729095753248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'], 139729095750896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['first'].weight, 139732776518096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B, 139729095739808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B['first'], 139729095751664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.base_layer, 139777661034864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout, 139729095751088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['first'], 139729095750560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v, 139729095737744) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A, 139729095739088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'], 139729095737408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['first'].weight, 139732776529856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B, 139729095738752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B['first'], 139729095741440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.base_layer, 139777661034912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout, 139729095743888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['first'], 139729095742544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k, 139777661034816) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.weight, 139777661101424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q, 139777661034672) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.weight, 139777661102784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.processor, 139777661034576) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm, 139777661034240) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm, 139777661034384) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu, 139777661034288) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear, 139729094571392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A, 139729094570528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'], 139729094562320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['first'].weight, 139732776724224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B, 139729094573456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B['first'], 139729094564624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.base_layer, 139777661034336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout, 139729094569712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['first'], 139729094570288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp, 139777661034480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp, 139729094566400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A, 139729094566304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'], 139729094558048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['first'].weight, 139732776717104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B, 139729094565056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B['first'], 139729094559296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.base_layer, 139777661034432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout, 139729094564720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['first'], 139729094566256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out, 139729094562416) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A, 139729094562704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['first'], 139729094565680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['first'].weight, 139732776518576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B, 139729094561936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B['first'], 139729094565104) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.base_layer, 139777661034528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout, 139729094560880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout['first'], 139729094559872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35], accessed_by=GetItemGuardAccessor(35) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35], 139777661034192) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn, 139777661035392) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k, 139729093790256) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A, 139729095007568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'], 139729095015968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['first'].weight, 139732776406528) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B, 139729095010448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B['first'], 139729095015824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.base_layer, 139777661035536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout, 139729093789728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['first'], 139729093792368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q, 139729093792176) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A, 139729093802736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'], 139729093788000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['first'].weight, 139732776403808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B, 139729093788768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B['first'], 139729093790208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.base_layer, 139777661035632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout, 139729093793280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['first'], 139729093795344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v, 139729095010880) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A, 139729095011984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'], 139729095005360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['first'].weight, 139732776409008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B, 139729095012224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B['first'], 139729095006848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.base_layer, 139777661035680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout, 139729095007280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['first'], 139729095009392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k, 139777661035584) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.weight, 139777661102544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q, 139777661035440) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.weight, 139777661103904) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.processor, 139777661035344) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm, 139777661035008) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm, 139777661035152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu, 139777661035056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear, 139729095750512) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A, 139729095745376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'], 139729093803456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['first'].weight, 139732776526256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B, 139729095747056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B['first'], 139729093794144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.base_layer, 139777661035104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout, 139729095742208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['first'], 139729095753632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp, 139777661035248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp, 139729093800768) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A, 139729093803024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'], 139729093798128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['first'].weight, 139732776518496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B, 139729093797216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B['first'], 139729093798896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.base_layer, 139777661035200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout, 139729093798416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['first'], 139729093795248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out, 139729093803072) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A, 139729093800720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['first'], 139729093794432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['first'].weight, 139732776409808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B, 139729093795872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B['first'], 139729093791984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.base_layer, 139777661035296) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout, 139729093801776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout['first'], 139729093799088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36], accessed_by=GetItemGuardAccessor(36) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36], 139777661034960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn, 139777661036160) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k, 139729098327424) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A, 139729098327088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'], 139729098330880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['first'].weight, 139732778353824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B, 139729098329632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B['first'], 139729098340048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.base_layer, 139777661036304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout, 139729098326176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['first'], 139729098326224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q, 139729098331264) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A, 139729098332080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'], 139729098332224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['first'].weight, 139732778363344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B, 139729098330016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B['first'], 139729098331216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.base_layer, 139777661036400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout, 139729098329872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['first'], 139729098331840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v, 139729098337696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A, 139729098789920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'], 139729098800816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['first'].weight, 139732778366544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B, 139729098799952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B['first'], 139729098796880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.base_layer, 139777661036448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout, 139729098335632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['first'], 139729098341152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k, 139777661036352) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.weight, 139777661103664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q, 139777661036208) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.weight, 139777661105024) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.processor, 139777661036112) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm, 139777661035776) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm, 139777661035920) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu, 139777661035824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear, 139729095003584) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A, 139729095006128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'], 139729095000944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['first'].weight, 139732776414128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B, 139729095004544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B['first'], 139729095000656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.base_layer, 139777661035872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout, 139729095004880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['first'], 139729095005264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp, 139777661036016) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp, 139729095000368) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A, 139729095001808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'], 139729095007136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['first'].weight, 139732776412048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B, 139729095003536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B['first'], 139729095010400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.base_layer, 139777661035968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout, 139729095002720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['first'], 139729095001712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out, 139729098339520) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A, 139729098338848) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['first'], 139729098336448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['first'].weight, 139732778353504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B, 139729098335536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B['first'], 139729098336544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.base_layer, 139777661036064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout, 139729098337840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout['first'], 139729098339040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37], accessed_by=GetItemGuardAccessor(37) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37], 139777661035728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn, 139777661036928) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k, 139729099849104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A, 139729099847808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'], 139729099838016) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['first'].weight, 139732778237472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B, 139729099848576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B['first'], 139729099844208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.base_layer, 139777661037072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout, 139729099846224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['first'], 139729099845216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_k.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_k.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_k.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q, 139729098795152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A, 139729099837968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'], 139729099839024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['first'].weight, 139732778222112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B, 139729099842528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B['first'], 139729099835424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.base_layer, 139777661037168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout, 139729099843824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['first'], 139729099842000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_q.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_q.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_q.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v, 139729099838208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A, 139729099841472) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'], 139729099837680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['first'].weight, 139732778230592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B, 139729099840608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B['first'], 139729099837824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.base_layer, 139777661037216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout, 139729099835856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['first'], 139729099837296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_v.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_v.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_v.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k, 139777661037120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.weight, 139777661104784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q, 139777661036976) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.weight, 139777661106144) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.processor, 139777661036880) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm, 139777661036544) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm, 139777661036688) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu, 139777661036592) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear, 139729098795440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A, 139729098793952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'], 139729098794816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['first'].weight, 139732778367904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B, 139729098796208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B['first'], 139729098796736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.base_layer, 139777661036640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout, 139729098798560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['first'], 139729098790208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].norm.linear.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].norm.linear.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].norm.linear.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp, 139777661036784) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp, 139729098789440) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A, 139729098787040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'], 139729098792752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['first'].weight, 139732778358784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B, 139729098788288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B['first'], 139729098792512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.base_layer, 139777661036736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout, 139729098794528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['first'], 139729098794624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_mlp.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].proj_mlp.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_mlp.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out, 139729098790736) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A, 139729098796544) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['first'], 139729098788432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['first'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['first']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['first'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['first'].weight, 139732778352704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B, 139729098797312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B['first'], 139729098788480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B['first'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.base_layer, 139777661036832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout, 139729098795584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['first'], accessed_by=GetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout['first'], 139729098796064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['first'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['first'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout['first'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_out.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].proj_out.scaling['first'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_out.use_dora) == 1 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.use_dora['first'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['img_ids'], accessed_by=DictGetItemGuardAccessor(img_ids) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['img_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[4096, 3], stride=[3, 1]) # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['img_ids'], '_dynamo_dynamic_indices') == False # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['txt_ids'], accessed_by=DictGetItemGuardAccessor(txt_ids) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['txt_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[512, 3], stride=[3, 1]) # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['txt_ids'], '_dynamo_dynamic_indices') == False # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['guidance'], accessed_by=DictGetItemGuardAccessor(guidance) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['guidance'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.float32, device=0, requires_grad=False, size=[1], stride=[1]) # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['guidance'], '_dynamo_dynamic_indices') == False # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['timestep'], accessed_by=DictGetItemGuardAccessor(timestep) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['timestep'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1], stride=[1]) # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['timestep'], '_dynamo_dynamic_indices') == False # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['return_dict'], accessed_by=DictGetItemGuardAccessor(return_dict) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- ID_MATCH: ___check_obj_id(L['return_dict'], 139786895605728) # if not return_dict: # diffusers/src/diffusers/models/transformers/transformer_flux.py:555 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['hidden_states'], accessed_by=DictGetItemGuardAccessor(hidden_states) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 4096, 64], stride=[262144, 64, 1]) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['pooled_projections'], accessed_by=DictGetItemGuardAccessor(pooled_projections) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['pooled_projections'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 768], stride=[768, 1]) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['pooled_projections'], '_dynamo_dynamic_indices') == False # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['encoder_hidden_states'], accessed_by=DictGetItemGuardAccessor(encoder_hidden_states) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- TENSOR_MATCH: check_tensor(L['encoder_hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 512, 4096], stride=[2097152, 4096, 1]) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_HASATTR: hasattr(L['encoder_hidden_states'], '_dynamo_dynamic_indices') == False # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['joint_attention_kwargs'], accessed_by=DictGetItemGuardAccessor(joint_attention_kwargs) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- ID_MATCH: ___check_obj_id(L['joint_attention_kwargs'], 139786895690912) # if joint_attention_kwargs is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['controlnet_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_block_samples) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_block_samples'], 139786895690912) # if controlnet_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:502 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=L['controlnet_single_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_single_block_samples) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_single_block_samples'], 139786895690912) # if controlnet_single_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:538 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | +- GuardManager: source=G, accessed_by=GlobalsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['torch'], accessed_by=DictGetItemGuardAccessor(torch) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['torch'], 139786870267216) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['torch'].cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['torch'].cat, 139786867250080) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['torch'].float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- EQUALS_MATCH: G['torch'].float16 == torch.float16 # if encoder_hidden_states.dtype == torch.float16: # diffusers/src/diffusers/models/transformers/transformer_flux.py:200 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['USE_PEFT_BACKEND'], accessed_by=DictGetItemGuardAccessor(USE_PEFT_BACKEND) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['USE_PEFT_BACKEND'], 139786895605696) # if USE_PEFT_BACKEND: # diffusers/src/diffusers/models/transformers/transformer_flux.py:434 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['scale_lora_layers'], accessed_by=DictGetItemGuardAccessor(scale_lora_layers) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['scale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['scale_lora_layers'].__code__, 139780914343312) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['unscale_lora_layers'], accessed_by=DictGetItemGuardAccessor(unscale_lora_layers) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['unscale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['unscale_lora_layers'].__code__, 139780914343488) # unscale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:553 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__builtins_dict___2'], accessed_by=DictGetItemGuardAccessor(__builtins_dict___2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['int'], accessed_by=DictGetItemGuardAccessor(int) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['int'], 139786895673536) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['len'], accessed_by=DictGetItemGuardAccessor(len) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['len'], 139786873098768) # assert len(timesteps.shape) == 1, "Timesteps should be a 1d-array" # diffusers/src/diffusers/models/embeddings.py:54 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['set'], accessed_by=DictGetItemGuardAccessor(set) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['set'], 139786895697184) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['str'], accessed_by=DictGetItemGuardAccessor(str) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['str'], 139786895715456) # if isinstance(self.active_adapter, str): # peft/tuners/tuners_utils.py:530 in active_adapters V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['bool'], accessed_by=DictGetItemGuardAccessor(bool) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['bool'], 139786895605760) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['range'], accessed_by=DictGetItemGuardAccessor(range) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['range'], 139786895693664) # for i in range(n_axes): # diffusers/src/diffusers/models/embeddings.py:696 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['enumerate'], accessed_by=DictGetItemGuardAccessor(enumerate) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['enumerate'], 139786895625344) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__builtins_dict___2']['isinstance'], accessed_by=DictGetItemGuardAccessor(isinstance) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___2']['isinstance'], 139786873098448) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'], accessed_by=DictGetItemGuardAccessor(__import_peft_dot_tuners_dot_tuners_utils) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'], 139781156739152) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, accessed_by=GetAttrGuardAccessor(BaseTunerLayer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, 94650012074624) # from peft.tuners.tuners_utils import BaseTunerLayer # diffusers/src/diffusers/utils/peft_utils.py:113 in scale_lora_layers V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention'], 139780514175456) # if len(args) > 0 or kwargs.get("scale", None) is not None: # diffusers/src/diffusers/models/attention.py:1197 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_embeddings) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'], 139780515771264) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np, accessed_by=GetAttrGuardAccessor(np) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np, 139786867275984) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, accessed_by=GetAttrGuardAccessor(ndarray) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, 139783921183872) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math, accessed_by=GetAttrGuardAccessor(math) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math, 139786870403888) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math.log, accessed_by=GetAttrGuardAccessor(log) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math.log, 139786870406768) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch, 139786870267216) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, 139786867250080) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, accessed_by=GetAttrGuardAccessor(cos) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, 139786867251280) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, accessed_by=GetAttrGuardAccessor(exp) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, 139786867252640) # emb = torch.exp(exponent) # diffusers/src/diffusers/models/embeddings.py:62 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, accessed_by=GetAttrGuardAccessor(sin) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, 139786867261040) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, accessed_by=GetAttrGuardAccessor(outer) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, 139786867273104) # freqs = torch.outer(pos, freqs) # type: ignore # [S, D/2] # diffusers/src/diffusers/models/embeddings.py:616 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, accessed_by=GetAttrGuardAccessor(stack) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, 139786867198048) # x_rotated = torch.stack([-x_imag, x_real], dim=-1).flatten(3) # diffusers/src/diffusers/models/embeddings.py:662 in apply_rotary_emb V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, accessed_by=GetAttrGuardAccessor(arange) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, 139786867122368) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float32 == torch.float32 # start=0, end=half_dim, dtype=torch.float32, device=timesteps.device # diffusers/src/diffusers/models/embeddings.py:58 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float64, accessed_by=GetAttrGuardAccessor(float64) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float64 == torch.float64 # freqs_dtype = torch.float32 if is_mps else torch.float64 # diffusers/src/diffusers/models/embeddings.py:695 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=GetAttrGuardAccessor(apply_rotary_emb) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, 139780515561872) # from .embeddings import apply_rotary_emb # diffusers/src/diffusers/models/attention_processor.py:1760 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], 139786895605696) # if use_real: # diffusers/src/diffusers/models/embeddings.py:653 in apply_rotary_emb V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1], accessed_by=GetItemGuardAccessor(1) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1] == -1 # if use_real_unbind_dim == -1: # diffusers/src/diffusers/models/embeddings.py:659 in apply_rotary_emb V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=GetAttrGuardAccessor(get_timestep_embedding) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, 139780515482768) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3] == 10000 # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=GetAttrGuardAccessor(get_1d_rotary_pos_embed) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, 139780515495616) # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=FuncDefaultsGuardAccessor V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0] == 10000.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2], accessed_by=GetItemGuardAccessor(2) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2] == 1.0 # 1.0 # diffusers/src/diffusers/models/embeddings.py:612 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3] == 1.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'], accessed_by=DictGetItemGuardAccessor(__import_torch_dot_nn_dot_modules_dot_module) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'], 139781214054384) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch, 139786870267216) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, accessed_by=GetAttrGuardAccessor(_C) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, 139786866637808) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, accessed_by=GetAttrGuardAccessor(_get_tracing_state) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, 139781219067088) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_pre_hooks) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_activations) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'], 139780515377728) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F, 139781213067664) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F.gelu, accessed_by=GetAttrGuardAccessor(gelu) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F.gelu, 139781219449200) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_normalization) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'], 139780513909392) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch, 139786870267216) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, accessed_by=GetAttrGuardAccessor(chunk) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, 139786867250240) # scale, shift = torch.chunk(emb, 2, dim=1) # diffusers/src/diffusers/models/normalization.py:305 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, accessed_by=GetAttrGuardAccessor(rsqrt) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, 139786867196688) # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float16 == torch.float16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float32 == torch.float32 # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16, accessed_by=GetAttrGuardAccessor(bfloat16) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16 == torch.bfloat16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention_processor) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'], 139780515379008) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F, 139781213067664) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, accessed_by=GetAttrGuardAccessor(scaled_dot_product_attention) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, 139781219601200) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch, 139786870267216) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, 139786867250080) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect, accessed_by=GetAttrGuardAccessor(inspect) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect, 139786871160896) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature, accessed_by=GetAttrGuardAccessor(signature) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, 139786868738336) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:13:53.735000 139786888062784 torch/_dynamo/guards.py:2148] [0/1] [__guards] V0909 15:14:40.363000 139786888062784 torch/_dynamo/guards.py:2611] [0/2] [__recompiles] Recompiling function forward in /home/sayak/diffusers/src/diffusers/models/transformers/transformer_flux.py:388 V0909 15:14:40.363000 139786888062784 torch/_dynamo/guards.py:2611] [0/2] [__recompiles] triggered by the following guard failure(s): V0909 15:14:40.363000 139786888062784 torch/_dynamo/guards.py:2611] [0/2] [__recompiles] - len(L['self'].transformer_blocks[0].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:14:40.363000 139786888062784 torch/_dynamo/guards.py:2611] [0/2] [__recompiles] - ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139777664822256) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward W0909 15:17:02.918000 139786888062784 torch/fx/experimental/symbolic_shapes.py:4449] [0/2] xindex is not in var_ranges, defaulting to unknown range. V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2169] [0/2] [__guards] GUARDS: V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] TREE_GUARD_MANAGER: V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] +- RootGuardManager V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- DEFAULT_DEVICE: utils_device.CURRENT_DEVICE == None # _dynamo/output_graph.py:460 in init_ambient_guards V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GLOBAL_STATE: ___check_global_state() V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['self'], accessed_by=DictGetItemGuardAccessor(self) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- ID_MATCH: ___check_obj_id(L['self'], 139777664825136) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=L['self'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=L['self'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(L['self'].training, 139786895605728) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=L['self']._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].norm_out, accessed_by=DictGetItemGuardAccessor(norm_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out, 139777661036496) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].norm_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].norm_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.training, 139786895605728) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm, 139777661037408) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu, 139777661037312) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear, 139777661037360) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out, 139777661037264) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out.training, 139786895605728) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].pos_embed, accessed_by=DictGetItemGuardAccessor(pos_embed) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed, 139777664824800) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].pos_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].pos_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed.training, 139786895605728) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim, accessed_by=DictGetItemGuardAccessor(axes_dim) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].pos_embed.axes_dim, 139786895700800) # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- LENGTH_CHECK: len(L['self'].pos_embed.axes_dim) == 3 # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[0], accessed_by=TupleGetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[0] == 16 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[1], accessed_by=TupleGetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[1] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[2], accessed_by=TupleGetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[2] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].x_embedder, accessed_by=DictGetItemGuardAccessor(x_embedder) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder, 139777664824032) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].x_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].x_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder.training, 139786895605728) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].time_text_embed, accessed_by=DictGetItemGuardAccessor(time_text_embed) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed, 139777664824848) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].time_text_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.training, 139786895605728) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj, accessed_by=DictGetItemGuardAccessor(time_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj, 139777664824752) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.training, 139786895605728) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.scale, accessed_by=DictGetItemGuardAccessor(scale) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.scale == 1 # scale=self.scale, # diffusers/src/diffusers/models/embeddings.py:769 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.num_channels, accessed_by=DictGetItemGuardAccessor(num_channels) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.num_channels == 256 # self.num_channels, # diffusers/src/diffusers/models/embeddings.py:766 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.flip_sin_to_cos, accessed_by=DictGetItemGuardAccessor(flip_sin_to_cos) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.flip_sin_to_cos, 139786895605696) # flip_sin_to_cos=self.flip_sin_to_cos, # diffusers/src/diffusers/models/embeddings.py:767 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.downscale_freq_shift, accessed_by=DictGetItemGuardAccessor(downscale_freq_shift) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.downscale_freq_shift == 0 # downscale_freq_shift=self.downscale_freq_shift, # diffusers/src/diffusers/models/embeddings.py:768 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder, accessed_by=DictGetItemGuardAccessor(text_embedder) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder, 139777664824416) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.text_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.training, 139786895605728) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1, accessed_by=DictGetItemGuardAccessor(act_1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1, 139777664824128) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1.training, 139786895605728) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1, 139777664824176) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1.training, 139786895605728) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2, 139777664824080) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2.training, 139786895605728) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder, accessed_by=DictGetItemGuardAccessor(guidance_embedder) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder, 139777664824560) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.guidance_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.training, 139786895605728) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1, 139777664824512) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2, 139777664824464) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder, accessed_by=DictGetItemGuardAccessor(timestep_embedder) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder, 139777664824704) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.timestep_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.training, 139786895605728) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1, 139777664824656) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2, 139777664824608) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].context_embedder, accessed_by=DictGetItemGuardAccessor(context_embedder) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder, 139777664824320) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].context_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].context_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder.training, 139786895605728) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].transformer_blocks, accessed_by=DictGetItemGuardAccessor(transformer_blocks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks, 139777664823936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0], 139777664823984) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff, 139777664822544) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net, 139777664822304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0], 139777664822352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139728980164176) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A, 139728980157744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'], 139724965838416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['second'].weight, 139722263474832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B, 139728980152176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['second'], 139724965834144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.base_layer, 139777664822256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout, 139728980150064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['second'], 139724965835680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1], 139777664822208) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2], 139728980164416) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A, 139728979916928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['second'], 139724965834288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['second'].weight, 139722263475952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B, 139728979915584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B['second'], 139724965838752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].base_layer, 139777664822160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout, 139728979906896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout['second'], 139724965834384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn, 139777664823360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k, 139728979952400) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A, 139728979958400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['second'], 139724965825936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['second'].weight, 139722263470752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B, 139728979965600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B['second'], 139724965827808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.base_layer, 139777664823168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout, 139728979953168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout['second'], 139724965828768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q, 139728980350656) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A, 139728980357280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['second'], 139724965825648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['second'].weight, 139722265039936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B, 139728980353008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B['second'], 139724965827136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.base_layer, 139777664823072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout, 139728980358864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout['second'], 139724965828960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v, 139728979954416) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A, 139728979960560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['second'], 139724965832992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['second'].weight, 139722263471312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B, 139728979958352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B['second'], 139724965836544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.base_layer, 139777664822976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout, 139728979964784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout['second'], 139724965838512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k, 139777664823120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.weight, 139777659850240) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q, 139777664823216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.weight, 139777660412416) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out, 139777664822784) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0], 139728980312704) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A, 139728980320144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'], 139724965823296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['second'].weight, 139722263473712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B, 139728980322880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B['second'], 139724965829200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].base_layer, 139777664822736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout, 139728980320048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['second'], 139724965828336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1], 139777664822688) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj, 139728980321920) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A, 139728980321872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'], 139724965831552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['second'].weight, 139722263471952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B, 139728980322112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['second'], 139724965832416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.base_layer, 139777664822928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout, 139728980321680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['second'], 139724965832176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj, 139728980317456) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A, 139728980318704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'], 139724965832512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['second'].weight, 139722263473392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B, 139728980328400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['second'], 139724965829152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.base_layer, 139777664822832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout, 139728980318800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['second'], 139724965838320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj, 139728980327920) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A, 139728980320768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'], 139724965829344) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['second'].weight, 139722263472672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B, 139728980317024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['second'], 139724965826368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.base_layer, 139777664822880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout, 139728980316016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['second'], 139724965829680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out, 139728978514016) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A, 139728978525632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'], 139724965835152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['second'].weight, 139722263474112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B, 139728980148576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B['second'], 139724965834768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.base_layer, 139777664822640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout, 139728978523904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['second'], 139724965834816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k, 139777664822496) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.weight, 139777664182976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q, 139777664822592) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.weight, 139777664182016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.processor, 139777664823408) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1, 139777664823840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm, 139777664823648) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu, 139777664823744) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear, 139728989471024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A, 139728989245152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['second'], 139724965837216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['second'].weight, 139722263728176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B, 139728966627440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B['second'], 139724965836448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.base_layer, 139777664823696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout, 139728989244624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout['second'], 139724965835248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear._active_adapter, 139786895671040) # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].norm1.linear._active_adapter) == 1 # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._active_adapter[0], accessed_by=ListGetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1.linear._active_adapter[0] == 'second' # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2, 139777664822448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context, 139777664822112) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net, 139777664821968) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0], 139777664822016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj, 139728979908240) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A, 139728979911984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'], 139724965826944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['second'].weight, 139722263476752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B, 139728979908624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['second'], 139724965826032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer, 139777664821920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout, 139728979903488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['second'], 139724965827040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1], 139777664821824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2], 139728978644320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A, 139728978645328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'], 139724965824496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['second'].weight, 139722263477312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B, 139728978644800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B['second'], 139724965823824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].base_layer, 139777664821776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout, 139728978644176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['second'], 139724965823728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context, 139777664823600) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm, 139777664823456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu, 139777664823552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear, 139728978846736) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A, 139728978842752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'], 139724965829584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['second'].weight, 139722265097088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B, 139728978844096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B['second'], 139724965837600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.base_layer, 139777664823504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout, 139728978850096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['second'], 139724965835440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context, 139777664822400) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1], 139777664823888) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff, 139777664820480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net, 139777664820240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0], 139777664820288) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj, 139728979049056) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A, 139728979048768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'], 139724965836352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['second'].weight, 139722263930144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B, 139728979036480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['second'], 139724965836256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.base_layer, 139777664820192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout, 139728979036816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['second'], 139724965836304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1], 139777664820144) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2], 139728978900976) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A, 139728978901648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['second'], 139724965831792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['second'].weight, 139722263930304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B, 139728978902032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B['second'], 139724965832032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].base_layer, 139777664820096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout, 139728978892768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout['second'], 139724965829296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn, 139777664821248) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k, 139728978656416) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A, 139728978970128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['second'], 139724965835632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['second'].weight, 139722263467552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B, 139728978969408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B['second'], 139724965835536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.base_layer, 139777664821104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout, 139728978970272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout['second'], 139724965835584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q, 139728978650272) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A, 139728978649456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['second'], 139724965837792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['second'].weight, 139722263468832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B, 139728978643168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B['second'], 139724965829440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.base_layer, 139777664821008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout, 139728978650368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout['second'], 139724965837408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v, 139728978974208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A, 139728978972288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['second'], 139724965824208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['second'].weight, 139722263463712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B, 139728978972192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B['second'], 139724965825168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.base_layer, 139777664820912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout, 139728978975792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout['second'], 139724965823440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k, 139777664821056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.weight, 139777660113664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q, 139777664821152) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.weight, 139777659965328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out, 139777664820720) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0], 139728980484416) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A, 139728980488928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'], 139724965822768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['second'].weight, 139722263928544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B, 139728980483696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B['second'], 139724965824976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].base_layer, 139777664820672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout, 139728980488496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['second'], 139724965822624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1], 139777664820624) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj, 139728978977424) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A, 139728978798256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'], 139724965830208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['second'].weight, 139722263926944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B, 139728978795856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['second'], 139724965831984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.base_layer, 139777664820864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout, 139728978974592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['second'], 139724965825840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj, 139728980442464) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A, 139728980489888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'], 139724965822864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['second'].weight, 139722263927984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B, 139728980482064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['second'], 139724965830256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.base_layer, 139777664820768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout, 139728980490128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['second'], 139724965823344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj, 139728989471600) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A, 139728980438096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'], 139724965831600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['second'].weight, 139722263929424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B, 139728980442896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['second'], 139724965838656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.base_layer, 139777664820816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout, 139728980438720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['second'], 139724965837168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out, 139728980489408) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A, 139728980490896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'], 139724965823872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['second'].weight, 139722263928944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B, 139728980476928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B['second'], 139724965824016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.base_layer, 139777664820576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout, 139728980491664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['second'], 139724965823968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k, 139777664820432) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.weight, 139777659965408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q, 139777664820528) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.weight, 139777659965488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.processor, 139777664821296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1, 139777664821728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm, 139777664821584) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu, 139777664821680) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear, 139728978654064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A, 139728978653872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['second'], 139724965833520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['second'].weight, 139722263475792) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B, 139728978652672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B['second'], 139724965835776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.base_layer, 139777664821632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout, 139728978654160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout['second'], 139724965833568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2, 139777664820384) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context, 139777664820048) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net, 139777664819904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0], 139777664819952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj, 139728978901744) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A, 139728978899200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'], 139724965832272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['second'].weight, 139722263930944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B, 139728978900880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['second'], 139724965831408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer, 139777664819856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout, 139728978902416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['second'], 139724965832224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1], 139777664819760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2], 139728978888448) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A, 139728978898384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'], 139724965823632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['second'].weight, 139722263931264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B, 139728978900064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B['second'], 139724965823248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].base_layer, 139777664819712) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout, 139728978901264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['second'], 139724965823584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context, 139777664821536) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm, 139777664821344) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu, 139777664821440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear, 139728978647152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A, 139728978647008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'], 139724965831504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['second'].weight, 139722263466032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B, 139728978644464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B['second'], 139724965832848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.base_layer, 139777664821392) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout, 139728978646624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['second'], 139724965831696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context, 139777664820336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2], 139777664822064) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff, 139777664818416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net, 139777664818176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0], 139777664818224) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj, 139728980519488) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A, 139728980519392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'], 139724964305968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['second'].weight, 139722263936464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B, 139728980519872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['second'], 139724964299536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.base_layer, 139777664818128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout, 139728980519536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['second'], 139724964309904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1], 139777664818080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2], 139728980516320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A, 139728980524576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['second'], 139724964299776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['second'].weight, 139722263936784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B, 139728980522800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B['second'], 139724964310912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].base_layer, 139777664818032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout, 139728980517472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout['second'], 139724964299392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn, 139777664819184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k, 139728977639600) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A, 139728977639744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['second'], 139724964311488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['second'].weight, 139722263932704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B, 139728977639696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B['second'], 139724964311008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.base_layer, 139777664819040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout, 139728977638784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout['second'], 139724964311392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q, 139728978789040) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A, 139728977625488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['second'], 139724964308512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['second'].weight, 139722263932304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B, 139728977625152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B['second'], 139724964306208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.base_layer, 139777664818944) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout, 139728977632256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout['second'], 139724964308320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v, 139728977632784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A, 139728977637488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['second'], 139724964314944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['second'].weight, 139722263933504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B, 139728977626256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B['second'], 139724964314320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.base_layer, 139777664818848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout, 139728977634896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout['second'], 139724964314368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k, 139777664818992) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.weight, 139777664567632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q, 139777664819088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.weight, 139777664569792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out, 139777664818656) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0], 139728979413168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A, 139728979426464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'], 139724964301312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['second'].weight, 139722263935344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B, 139728979426368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B['second'], 139724964301792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].base_layer, 139777664818608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout, 139728979422144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['second'], 139724964301840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1], 139777664818560) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj, 139728978452848) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A, 139728978454480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'], 139724964309232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['second'].weight, 139722263933984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B, 139728978456976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['second'], 139724964309616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.base_layer, 139777664818800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout, 139728978444448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['second'], 139724964304000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj, 139728978506368) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A, 139728978499120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'], 139724964300448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['second'].weight, 139722263934944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B, 139728978499360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['second'], 139724964300304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.base_layer, 139777664818704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout, 139728978500992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['second'], 139724964300976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj, 139728978458032) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A, 139728978458608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'], 139724964302896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['second'].weight, 139722263934224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B, 139728978456064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['second'], 139724964303616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.base_layer, 139777664818752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout, 139728978450208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['second'], 139724964303664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out, 139728979423344) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A, 139728979419120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'], 139724964314272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['second'].weight, 139722263935824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B, 139728979420656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B['second'], 139724964299008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.base_layer, 139777664818512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout, 139728979421040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['second'], 139724964300064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k, 139777664818368) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.weight, 139777664568672) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q, 139777664818464) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.weight, 139777664567552) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.processor, 139777664819232) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1, 139777664819664) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm, 139777664819520) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu, 139777664819616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear, 139728978887296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A, 139728978887584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['second'], 139724965826560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['second'].weight, 139722263931824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B, 139728978887824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B['second'], 139724965826224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.base_layer, 139777664819568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout, 139728978897568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout['second'], 139724965828144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2, 139777664818320) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context, 139777664817984) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net, 139777664817840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0], 139777664817888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj, 139728980516272) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A, 139728980512000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'], 139724964308992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['second'].weight, 139722263920784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B, 139728979094896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['second'], 139724964305824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer, 139777664817792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout, 139728980522944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['second'], 139724964311056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1], 139777664817696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2], 139728979089472) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A, 139728979092208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'], 139724964309424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['second'].weight, 139722263921424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B, 139728979097440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B['second'], 139724964309568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].base_layer, 139777664817648) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout, 139728979097632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['second'], 139724964304480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context, 139777664819472) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm, 139777664819280) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu, 139777664819376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear, 139728978901024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A, 139728978900784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'], 139724964307216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['second'].weight, 139722263932224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B, 139728978900256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B['second'], 139724964307312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.base_layer, 139777664819328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout, 139728978888160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['second'], 139724965826800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context, 139777664818272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3], 139777664820000) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff, 139777664816352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net, 139777664816112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0], 139777664816160) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj, 139728977664448) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A, 139728977662576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'], 139724964305488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['second'].weight, 139722358886128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B, 139728977663920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['second'], 139724964304240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.base_layer, 139777664816064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout, 139728977665504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['second'], 139724964304384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1], 139777664816016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2], 139728977672992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A, 139728977665408) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['second'], 139724964299824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['second'].weight, 139722357098512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B, 139728977664592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B['second'], 139724964304528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].base_layer, 139777664815968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout, 139728977665168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout['second'], 139724964299872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn, 139777664817120) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k, 139728979281472) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A, 139728979293040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['second'], 139724964300544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['second'].weight, 139722347853152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B, 139728979293520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B['second'], 139724964302512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.base_layer, 139777664816976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout, 139728979280608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout['second'], 139724964300640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q, 139728979284784) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A, 139728979283584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['second'], 139724964303568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['second'].weight, 139722347839712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B, 139728979283536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B['second'], 139724964303472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.base_layer, 139777664816880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout, 139728979286608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout['second'], 139724964303520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v, 139728979144432) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A, 139728979137280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['second'], 139724964307120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['second'].weight, 139722347853552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B, 139728979137616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B['second'], 139724964307024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.base_layer, 139777664816784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout, 139728979136608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout['second'], 139724964307072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k, 139777664816928) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.weight, 139777659970608) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q, 139777664817024) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.weight, 139777664565792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out, 139777664816592) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0], 139728978811136) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A, 139728978811184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'], 139724964299632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['second'].weight, 139722342120512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B, 139728978817136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B['second'], 139724964313840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].base_layer, 139777664816544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout, 139728978820496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['second'], 139724964307744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1], 139777664816496) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj, 139728979138432) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A, 139728979135456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'], 139724964313696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['second'].weight, 139722347854352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B, 139728979138480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['second'], 139724964301696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.base_layer, 139777664816736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout, 139728979135216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['second'], 139724964313648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj, 139728979142224) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A, 139728978807968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'], 139724964302176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['second'].weight, 139722342120032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B, 139728978820160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['second'], 139724964303376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.base_layer, 139777664816640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout, 139728978819824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['second'], 139724964301744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj, 139728979142320) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A, 139728979133200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'], 139724964306352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['second'].weight, 139722349072912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B, 139728979136656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['second'], 139724964306928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.base_layer, 139777664816688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout, 139728979139776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['second'], 139724964306256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out, 139728978810896) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A, 139728978815984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'], 139724964313936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['second'].weight, 139722342112672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B, 139728978816224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B['second'], 139724964310432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.base_layer, 139777664816448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout, 139728978819920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['second'], 139724964313888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k, 139777664816304) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.weight, 139777659970448) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q, 139777664816400) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.weight, 139777659970528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.processor, 139777664817168) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1, 139777664817600) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm, 139777664817456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu, 139777664817552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear, 139728978751648) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A, 139728978754192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['second'], 139724964309328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['second'].weight, 139722263921744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B, 139728978746656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B['second'], 139724964304672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.base_layer, 139777664817504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout, 139728978751696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout['second'], 139724964309520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2, 139777664816256) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context, 139777664815920) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net, 139777664815776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0], 139777664815824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj, 139728979338592) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A, 139728979342672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'], 139724964312448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['second'].weight, 139722341421424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B, 139728979342000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['second'], 139724964298912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer, 139777664815728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout, 139728979339600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['second'], 139724964312304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1], 139777664815632) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2], 139728977447568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A, 139728979210256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'], 139724964311152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['second'].weight, 139722341417264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B, 139728979212656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B['second'], 139724964311776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].base_layer, 139777664815584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout, 139728977456496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['second'], 139724964311632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context, 139777664817408) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm, 139777664817216) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu, 139777664817312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear, 139728979286992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A, 139728979286800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'], 139724964308560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['second'].weight, 139722622530512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B, 139728979284832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B['second'], 139724964314656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.base_layer, 139777664817264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout, 139728979287088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['second'], 139724964308176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context, 139777664816208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4], 139777664817936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff, 139777664814288) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net, 139777664814048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0], 139777664814096) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj, 139728977554704) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A, 139728977558832) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'], 139724964452368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['second'].weight, 139722341428304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B, 139728978036816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['second'], 139724964449248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.base_layer, 139777664814000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout, 139728977558592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['second'], 139724964448816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1], 139777664813952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2], 139728978048528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A, 139728978046800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['second'], 139724964450880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['second'].weight, 139722341428704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B, 139728978035664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B['second'], 139724964451984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].base_layer, 139777664813904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout, 139728978037584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout['second'], 139724964452176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn, 139777664815056) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k, 139728978900544) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A, 139728977775632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['second'], 139724964312640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['second'].weight, 139722341423824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B, 139728977773328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B['second'], 139724964314176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.base_layer, 139777664814912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout, 139728977775248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout['second'], 139724964313984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q, 139728979210304) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A, 139728979375248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['second'], 139724964311968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['second'].weight, 139722341423264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B, 139728979362960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B['second'], 139724964309472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.base_layer, 139777664814816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout, 139728979373088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout['second'], 139724964311920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v, 139728977774192) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A, 139728977773136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['second'], 139724964300928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['second'].weight, 139722341424384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B, 139728977773184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B['second'], 139724964313216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.base_layer, 139777664814720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout, 139728977774528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout['second'], 139724964301072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k, 139777664814864) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.weight, 139777664570512) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q, 139777664814960) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.weight, 139777664571712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out, 139777664814528) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0], 139728979356560) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A, 139728979356416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'], 139724964460912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['second'].weight, 139722341426224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B, 139728979361408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B['second'], 139724964456496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].base_layer, 139777664814480) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout, 139728979348784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['second'], 139724964454624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1], 139777664814432) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj, 139728977779760) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A, 139728977788064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'], 139730416024256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['second'].weight, 139722341424864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B, 139728977780576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['second'], 139730416015568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.base_layer, 139777664814672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout, 139728977788784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['second'], 139724964302224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj, 139728980379200) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A, 139728980390816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'], 139724964461440) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['second'].weight, 139722341425744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B, 139728980380976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['second'], 139724964453376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.base_layer, 139777664814576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout, 139728980381888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['second'], 139730416013456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj, 139728977773808) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A, 139728980384576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'], 139730416016720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['second'].weight, 139722341425344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B, 139728980389328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['second'], 139730416009376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.base_layer, 139777664814624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout, 139728980385968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['second'], 139730416020896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out, 139728977554176) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A, 139728977557488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'], 139724964452656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['second'].weight, 139722341427984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B, 139728977556720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B['second'], 139724964461776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.base_layer, 139777664814384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout, 139728977549616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['second'], 139724964460720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k, 139777664814240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.weight, 139777659971408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q, 139777664814336) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.weight, 139777664568832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.processor, 139777664815104) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1, 139777664815536) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm, 139777664815392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu, 139777664815488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear, 139728979208528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A, 139728979203584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['second'], 139724964300688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['second'].weight, 139722341417824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B, 139728979206608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B['second'], 139724964302992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.base_layer, 139777664815440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout, 139728979206416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout['second'], 139724964303904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2, 139777664814192) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context, 139777664813856) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net, 139777664813712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0], 139777664813760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj, 139728979365408) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A, 139728976457936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'], 139724964449776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['second'].weight, 139722341429104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B, 139728976456784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['second'], 139724964453904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer, 139777664813664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout, 139728976447712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['second'], 139724964450784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1], 139777664813568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2], 139728976450976) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A, 139728976460432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'], 139724964447664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['second'].weight, 139722341433184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B, 139728976451360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B['second'], 139724964447424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].base_layer, 139777664813520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout, 139728976451264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['second'], 139724964447376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context, 139777664815344) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm, 139777664815152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu, 139777664815248) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear, 139728979209344) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A, 139728979207856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'], 139724964310624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['second'].weight, 139722341422784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B, 139728979212560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B['second'], 139724964302704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.base_layer, 139777664815200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout, 139728979203008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['second'], 139724964310864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context, 139777664814144) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5], 139777664815872) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff, 139777664812224) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net, 139777664811984) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0], 139777664812032) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj, 139728978177056) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A, 139728979063520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'], 139724964448432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['second'].weight, 139722341486000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B, 139728979062752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['second'], 139724964458752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.base_layer, 139777664811936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout, 139728979054592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['second'], 139724964448384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1], 139777664811888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2], 139728979052672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A, 139728979058912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['second'], 139724964462496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['second'].weight, 139722341370752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B, 139728978307696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B['second'], 139724964457888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].base_layer, 139777664811840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout, 139728979052960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout['second'], 139724964457696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn, 139777664812992) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k, 139728977932304) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A, 139728977935808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['second'], 139724964454768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['second'].weight, 139722341482960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B, 139728977934464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B['second'], 139724964461488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.base_layer, 139777664812848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout, 139728977923232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout['second'], 139724964454720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q, 139728976378944) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A, 139728976373232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['second'], 139724964455920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['second'].weight, 139722341492240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B, 139728976377984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B['second'], 139724964446560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.base_layer, 139777664812752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout, 139728976377840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout['second'], 139724964452512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v, 139728977920784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A, 139728977926160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['second'], 139724964460480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['second'].weight, 139722341483360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B, 139728978168224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B['second'], 139724964457744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.base_layer, 139777664812656) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout, 139728977931776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout['second'], 139724964458416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k, 139777664812800) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.weight, 139781198830368) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q, 139777664812896) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.weight, 139777664567072) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out, 139777664812464) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0], 139728978179600) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A, 139728978178160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'], 139724964457120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['second'].weight, 139722341485040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B, 139728978176816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B['second'], 139724964457264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].base_layer, 139777664812416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout, 139728978173360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['second'], 139724964456928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1], 139777664812368) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj, 139728978181328) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A, 139728978181280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'], 139724964461584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['second'].weight, 139722341483760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B, 139728978180512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['second'], 139724964456880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.base_layer, 139777664812608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout, 139728978181376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['second'], 139724964458848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj, 139728978176864) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A, 139728978176480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'], 139724964454336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['second'].weight, 139722341484640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B, 139728978176048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['second'], 139724964449344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.base_layer, 139777664812512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout, 139728978173984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['second'], 139724964455152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj, 139728978175808) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A, 139728978174080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'], 139724964457456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['second'].weight, 139722341484240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B, 139728978169760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['second'], 139724964461920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.base_layer, 139777664812560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout, 139728978180032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['second'], 139724964457552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out, 139728978171728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A, 139728978171440) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'], 139724964459808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['second'].weight, 139722341485440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B, 139728978173648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B['second'], 139724964447808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.base_layer, 139777664812320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout, 139728978178016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['second'], 139724964447616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k, 139777664812176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.weight, 139781198834848) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q, 139777664812272) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.weight, 139777664184096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.processor, 139777664813040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1, 139777664813472) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm, 139777664813328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu, 139777664813424) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear, 139728977894256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A, 139728976365840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['second'], 139724964448960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['second'].weight, 139722341497520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B, 139728976367952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B['second'], 139724964460768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.base_layer, 139777664813376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout, 139728977894640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout['second'], 139724964452608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2, 139777664812128) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context, 139777664811792) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net, 139777664811648) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0], 139777664811696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj, 139728978312400) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A, 139728978302128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'], 139724964454144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['second'].weight, 139722341370272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B, 139728978306352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['second'], 139724964455344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer, 139777664811600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout, 139728978307552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['second'], 139724964449152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1], 139777664811504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2], 139728976484848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A, 139728976478656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'], 139724964461680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['second'].weight, 139722341368352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B, 139728976488352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B['second'], 139724964454048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].base_layer, 139777664811456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout, 139728976492384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['second'], 139724964461632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context, 139777664813280) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm, 139777664813088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu, 139777664813184) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear, 139728976368000) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A, 139728976366464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'], 139724964456448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['second'].weight, 139722341487840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B, 139728976366368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B['second'], 139724964458560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.base_layer, 139777664813136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout, 139728976367712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['second'], 139724964453856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context, 139777664812080) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6], 139777664813808) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff, 139777674394928) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net, 139777674394688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0], 139777674394736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj, 139728978388992) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A, 139728978390240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'], 139724964377216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['second'].weight, 139722341383952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B, 139728978390144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['second'], 139724964373328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.base_layer, 139777674394640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout, 139728978389664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['second'], 139724964451024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1], 139777674394592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2], 139728978395040) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A, 139728978381456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['second'], 139724964369920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['second'].weight, 139722341374032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B, 139728978385392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B['second'], 139724964377168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].base_layer, 139777674394544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout, 139728978392016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout['second'], 139724964373808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn, 139777674383792) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k, 139728977441888) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A, 139728977436224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['second'], 139724964456112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['second'].weight, 139722341375632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B, 139728977434160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B['second'], 139724964462112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.base_layer, 139777674395552) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout, 139728977443136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout['second'], 139724964456016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q, 139728978068288) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A, 139728978074768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['second'], 139724964450208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['second'].weight, 139722341375072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B, 139728978074912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B['second'], 139724964450304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.base_layer, 139777674395456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout, 139728978068624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout['second'], 139724964450256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v, 139728977959648) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A, 139728977958592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['second'], 139724964456544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['second'].weight, 139722341376032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B, 139728977959888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B['second'], 139724964456640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.base_layer, 139777674395360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout, 139728977964688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout['second'], 139724964455968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k, 139777674395504) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.weight, 139777664177376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q, 139777674395600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.weight, 139777659972208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out, 139777674395168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0], 139728977486000) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A, 139728977481536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'], 139724964462448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['second'].weight, 139722341383072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B, 139728978393744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B['second'], 139724964460336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].base_layer, 139777674395120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout, 139728977490512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['second'], 139724964458368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1], 139777674395072) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj, 139728977968144) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A, 139728977825792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'], 139724964446608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['second'].weight, 139722341381552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B, 139728977834000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['second'], 139724964446416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.base_layer, 139777674395312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout, 139728977960032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['second'], 139724964446272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj, 139728977485376) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A, 139728977485232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'], 139724964459952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['second'].weight, 139722341382512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B, 139728977484944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['second'], 139724964450928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.base_layer, 139777674395216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout, 139728977491136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['second'], 139724964460096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj, 139728977834288) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A, 139728977478512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'], 139724964450064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['second'].weight, 139722341382032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B, 139728977488304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['second'], 139724964448096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.base_layer, 139777674395264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout, 139728977832896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['second'], 139724964449632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out, 139728978385440) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A, 139728978378960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'], 139724964446800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['second'].weight, 139722341383552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B, 139728978381744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B['second'], 139724964446464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.base_layer, 139777674395024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout, 139728978385056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['second'], 139724964446896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k, 139777674394880) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.weight, 139777664570752) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q, 139777674394976) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.weight, 139777664177456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.processor, 139777674380624) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1, 139777664811408) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm, 139777664811264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu, 139777664811360) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear, 139728978080576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A, 139728978067760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['second'], 139724964460528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['second'].weight, 139722341369552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B, 139728978075152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B['second'], 139724964451168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.base_layer, 139777664811312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout, 139728978068144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout['second'], 139724964451600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2, 139777674394832) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context, 139777674394496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net, 139777674394352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0], 139777674394400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj, 139728978220448) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A, 139728978216320) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'], 139724964378992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['second'].weight, 139722341377392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B, 139728978215744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['second'], 139724964371408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer, 139777674394256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout, 139728978221168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['second'], 139724964373616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1], 139777665008928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2], 139728978230048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A, 139728978224336) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'], 139724964378800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['second'].weight, 139722341371392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B, 139728978025664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B['second'], 139724964377600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].base_layer, 139777665009312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout, 139728978230576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['second'], 139724964376304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context, 139777664811216) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm, 139777674384080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu, 139777664811120) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear, 139728978079040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A, 139728978078752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'], 139724964451408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['second'].weight, 139722341374512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B, 139728978083504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B['second'], 139724964447184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.base_layer, 139777664811072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout, 139728978067616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['second'], 139724964447472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context, 139777674394784) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7], 139777664811744) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff, 139777663769216) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net, 139777663770080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0], 139777663770656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj, 139728975703024) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A, 139728975697600) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'], 139724964379952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['second'].weight, 139722341555776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B, 139728975707776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['second'], 139724964377360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.base_layer, 139777663770416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout, 139728975695680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['second'], 139724964379904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1], 139777663769504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2], 139728975700336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A, 139728975701632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['second'], 139724964378512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['second'].weight, 139722341556176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B, 139728975702064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B['second'], 139724964371744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].base_layer, 139777663769648) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout, 139728975693472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout['second'], 139724964378032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn, 139777665009504) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k, 139728977295104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A, 139728976543904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['second'], 139724964365408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['second'].weight, 139722341555056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B, 139728976546928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B['second'], 139724964376976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.base_layer, 139780515283440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout, 139728977297264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout['second'], 139724964367856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q, 139728978033104) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A, 139728978025328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['second'], 139724964366464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['second'].weight, 139722341378032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B, 139728978026288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B['second'], 139724964375152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.base_layer, 139786872078544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout, 139728978031472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout['second'], 139724964374528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v, 139728976546496) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A, 139728976556912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['second'], 139724964368096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['second'].weight, 139722341548496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B, 139728976551536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B['second'], 139724964372368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.base_layer, 139780515283008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout, 139728976550192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout['second'], 139724964369536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k, 139786872288784) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.weight, 139777664175376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q, 139777676533120) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.weight, 139781198836928) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out, 139777663769456) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0], 139728975827376) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A, 139728975834624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'], 139724964372608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['second'].weight, 139722341550256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B, 139728975835248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B['second'], 139724964373136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].base_layer, 139777663769408) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout, 139728975830448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['second'], 139724964375968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1], 139777663769360) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj, 139728975832032) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A, 139728975831552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'], 139724964368576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['second'].weight, 139722341562896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B, 139728975830928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['second'], 139724964368240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.base_layer, 139780515287040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout, 139728975830880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['second'], 139724964370880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj, 139728975835632) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A, 139728975835152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'], 139724964377552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['second'].weight, 139722341549616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B, 139728975833088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['second'], 139724964376160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.base_layer, 139780515287088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout, 139728975834144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['second'], 139724964374624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj, 139728975836496) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A, 139728975836448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'], 139724964373760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['second'].weight, 139722341550736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B, 139728975833856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['second'], 139724964375824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.base_layer, 139780515287136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout, 139728975831696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['second'], 139724964377792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out, 139728975831168) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A, 139728975704368) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'], 139724964364928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['second'].weight, 139722341555296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B, 139728975707680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B['second'], 139724964367520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.base_layer, 139777663769312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout, 139728975823824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['second'], 139724964364640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k, 139777663769120) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.weight, 139777660418256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q, 139777663769264) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.weight, 139777664178256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.processor, 139777665009072) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1, 139777665009360) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm, 139777665009456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu, 139777665009408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear, 139728978027824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A, 139728978034640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['second'], 139724964367760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['second'].weight, 139722341373232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B, 139728978026576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B['second'], 139724964366080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.base_layer, 139777665008976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout, 139728978025904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout['second'], 139724964366032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2, 139777663769552) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context, 139777663769744) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net, 139777663769600) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0], 139777663769696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj, 139728975693808) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A, 139728975698320) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'], 139724964374192) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['second'].weight, 139722341556736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B, 139728975706336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['second'], 139724964374816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer, 139777663770176) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout, 139728975704656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['second'], 139724964374240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1], 139777663770128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2], 139728976639792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A, 139728976636240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'], 139724964367280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['second'].weight, 139722341557136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B, 139728976641472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B['second'], 139724964366416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].base_layer, 139777663770320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout, 139728976639360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['second'], 139724964365696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context, 139777665009120) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm, 139777665007728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu, 139777665009024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear, 139728978029936) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A, 139728978034064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'], 139724964364592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['second'].weight, 139722341372672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B, 139728978020000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B['second'], 139724964367904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.base_layer, 139777665007824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout, 139728978033776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['second'], 139724964367952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context, 139777663770752) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8], 139777674394448) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff, 139777663771472) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net, 139777663771712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0], 139777663771664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj, 139728975371312) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A, 139728975368768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'], 139729099843152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['second'].weight, 139722341628832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B, 139728975378800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['second'], 139729099838640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.base_layer, 139777663771760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout, 139728975374528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['second'], 139729099840320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1], 139777663771808) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2], 139728976625728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A, 139728975415328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['second'], 139729099840128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['second'].weight, 139722341621392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B, 139728975415040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B['second'], 139730415059776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].base_layer, 139777663771856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout, 139728975421328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout['second'], 139729099840416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn, 139777663770848) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k, 139728976642000) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A, 139728976637968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['second'], 139724964378416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['second'].weight, 139722341563536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B, 139728976636912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B['second'], 139724964372896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.base_layer, 139777663770800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout, 139728976627984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout['second'], 139724964372656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q, 139728976632832) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A, 139728976625872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['second'], 139724964371168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['second'].weight, 139722341558576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B, 139728976630720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B['second'], 139724964375008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.base_layer, 139777663770944) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout, 139728976637296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout['second'], 139724964378272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v, 139728978038448) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A, 139728975328112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['second'], 139724964371504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['second'].weight, 139722341563936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B, 139728975328256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B['second'], 139724964369200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.base_layer, 139777663771040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout, 139728975329216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout['second'], 139724964371840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k, 139777663770896) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.weight, 139777659939520) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q, 139777663770704) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.weight, 139777664578912) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out, 139777663771232) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0], 139728975372176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A, 139728975378704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'], 139729099833600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['second'].weight, 139722341625872) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B, 139728975369728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B['second'], 139729099835328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].base_layer, 139777663771280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout, 139728975376160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['second'], 139724964380576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1], 139777663771328) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj, 139728975325568) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A, 139728976724656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'], 139724964366512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['second'].weight, 139722341564336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B, 139728976729312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['second'], 139724964366848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.base_layer, 139777663771088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout, 139728975325904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['second'], 139724964366560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj, 139728976726048) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A, 139728976737520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'], 139724964376496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['second'].weight, 139722341622992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B, 139728976737472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['second'], 139724964370352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.base_layer, 139777663771184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout, 139728976733056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['second'], 139724964376112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj, 139728976730704) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A, 139728976740064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'], 139724964371120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['second'].weight, 139722341627632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B, 139728976740112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['second'], 139724964373232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.base_layer, 139777663771136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout, 139728976731040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['second'], 139724964378080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out, 139728975375248) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A, 139728975374192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'], 139729099841712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['second'].weight, 139722341626032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B, 139728975376016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B['second'], 139729099835232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.base_layer, 139777663771376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout, 139728975374048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['second'], 139729099833744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k, 139777663771520) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.weight, 139777660416816) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q, 139777663771424) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.weight, 139777664185616) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.processor, 139777663770512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1, 139777663769840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm, 139777663769984) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu, 139777663769888) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear, 139728976636384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A, 139728976638208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['second'], 139724964365792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['second'].weight, 139722341557616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B, 139728976638304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B['second'], 139724964365888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.base_layer, 139777663769936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout, 139728976636528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout['second'], 139724964365840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2, 139777663771568) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context, 139777663771904) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net, 139777663772048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0], 139777663772000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj, 139728975414992) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A, 139728975413936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'], 139730415059248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['second'].weight, 139722341614032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B, 139728975420080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['second'], 139730415075088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer, 139777663772096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout, 139728975429248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['second'], 139730415071776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1], 139777663772192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2], 139728976002496) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A, 139728976002592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'], 139726562735056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['second'].weight, 139722341614432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B, 139728976002928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B['second'], 139726562735728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].base_layer, 139777663772240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout, 139728976002112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['second'], 139726562735152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context, 139777663770560) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm, 139777663770032) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu, 139777663770224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear, 139728976635376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A, 139728976639888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'], 139724964375584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['second'].weight, 139722341558176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B, 139728976640752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B['second'], 139724964380480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.base_layer, 139777663770368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout, 139728976627936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['second'], 139724964380144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context, 139777663771616) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9], 139777663769792) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff, 139777663773536) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net, 139777663773776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0], 139777663773728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj, 139728975643552) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A, 139728975649696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'], 139726564095456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['second'].weight, 139722337018208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B, 139728975643696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['second'], 139726564084416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.base_layer, 139777663773824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout, 139728975645328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['second'], 139726564284816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1], 139777663773872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2], 139728975646576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A, 139728975647392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['second'], 139726564084608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['second'].weight, 139722337018768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B, 139728975646000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B['second'], 139726564436304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].base_layer, 139777663773920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout, 139728975646096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout['second'], 139726564084704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn, 139777663772768) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k, 139728977399040) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A, 139728977397984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['second'], 139726562735968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['second'].weight, 139722337010848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B, 139728977397264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B['second'], 139726562724304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.base_layer, 139777663772912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout, 139728977396064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout['second'], 139726562736208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q, 139728977410176) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A, 139728977405760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['second'], 139726562737360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['second'].weight, 139722341615632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B, 139728977407248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B['second'], 139726562737504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.base_layer, 139777663773008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout, 139728977409408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout['second'], 139726562737456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v, 139728977397312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A, 139728977399904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['second'], 139726562653712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['second'].weight, 139722337024448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B, 139728977397552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B['second'], 139726562642768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.base_layer, 139777663773104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout, 139728977409648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout['second'], 139726562653808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k, 139777663772960) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.weight, 139777664576272) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q, 139777663772864) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.weight, 139777664576352) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out, 139777663773296) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0], 139728975767792) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A, 139728975606592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'], 139726564122128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['second'].weight, 139722337017328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B, 139728975600592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B['second'], 139726564127312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].base_layer, 139777663773344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout, 139728975597616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['second'], 139726562556176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1], 139777663773392) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj, 139728975547024) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A, 139728975556864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'], 139726564474352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['second'].weight, 139722337019088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B, 139728975549760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['second'], 139726564474064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.base_layer, 139777663773152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout, 139728975549280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['second'], 139726562643152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj, 139728975769184) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A, 139728975771920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'], 139726562545424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['second'].weight, 139722337016928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B, 139728975768608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['second'], 139726562545136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.base_layer, 139777663773248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout, 139728975757760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['second'], 139726564561424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj, 139728975769328) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A, 139728975759536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'], 139726564571984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['second'].weight, 139722337016528) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B, 139728975767312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['second'], 139726564571648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.base_layer, 139777663773200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout, 139728975768464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['second'], 139726564463792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out, 139728975602704) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A, 139728975647584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'], 139726564192128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['second'].weight, 139722337017728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B, 139728975647488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B['second'], 139726564187328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.base_layer, 139777663773440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout, 139728975647728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['second'], 139726564122464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k, 139777663773584) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.weight, 139777664177936) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q, 139777663773488) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.weight, 139777664576992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.processor, 139777663772720) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1, 139777663772288) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm, 139777663772432) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu, 139777663772336) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear, 139728977408160) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A, 139728977408112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['second'], 139726562735440) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['second'].weight, 139722341614832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B, 139728977407968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B['second'], 139726562735584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.base_layer, 139777663772384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout, 139728977408208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout['second'], 139726562735536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2, 139777663773632) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context, 139777663773968) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net, 139777663774112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0], 139777663774064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj, 139728975646960) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A, 139728975924768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'], 139726564436544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['second'].weight, 139722337010368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B, 139728975936864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['second'], 139726564338720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer, 139777663774160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout, 139728975647680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['second'], 139726564436640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1], 139777663774256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2], 139728975897952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A, 139728975900880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'], 139726564338672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['second'].weight, 139722337013968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B, 139728975899680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B['second'], 139726563997392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].base_layer, 139777663774304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout, 139728975898720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['second'], 139726564328304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context, 139777663772480) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm, 139777663772672) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu, 139777663772576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear, 139728977403168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A, 139728977405184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'], 139726562737168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['second'].weight, 139722341615232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B, 139728977407392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B['second'], 139726562737072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.base_layer, 139777663772624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout, 139728977405232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['second'], 139726562736880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context, 139777663773680) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10], 139777663771952) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff, 139777663775600) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net, 139777663775840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0], 139777663775792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj, 139728976051504) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A, 139728976046176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'], 139729098430576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['second'].weight, 139722338362336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B, 139728976049488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['second'], 139729098438448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.base_layer, 139777663775888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout, 139728976037632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['second'], 139729098429424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1], 139777663775936) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2], 139728976049872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A, 139728974655024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['second'], 139733030700896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['second'].weight, 139722338358256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B, 139728974655600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B['second'], 139729096092176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].base_layer, 139777663775984) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout, 139728974658480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout['second'], 139733030960672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn, 139777663774832) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k, 139728976191376) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A, 139728976190992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['second'], 139733023470064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['second'].weight, 139722337021808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B, 139728976192624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B['second'], 139733023461040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.base_layer, 139777663774976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout, 139728976191424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout['second'], 139733023463632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q, 139728975895216) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A, 139728975898432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['second'], 139733023467136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['second'].weight, 139722337021488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B, 139728976191280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B['second'], 139733023457536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.base_layer, 139777663775072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout, 139728975902272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout['second'], 139733023461328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v, 139728976199008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A, 139728976192576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['second'], 139733023473568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['second'].weight, 139722337022288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B, 139728976190896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B['second'], 139733023466032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.base_layer, 139777663775168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout, 139728976199152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout['second'], 139733023473520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k, 139777663775024) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.weight, 139777664181216) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q, 139777663774928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.weight, 139777664573792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out, 139777663775360) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0], 139728977387696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A, 139728977382080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'], 139729098429472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['second'].weight, 139722338355776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B, 139728977379872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B['second'], 139729098424912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].base_layer, 139777663775408) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout, 139728977380976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['second'], 139729098428464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1], 139777663775456) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj, 139728976188112) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A, 139728976199536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'], 139733021800384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['second'].weight, 139722338361616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B, 139728976197376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['second'], 139733021798656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.base_layer, 139777663775216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout, 139728976184032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['second'], 139733023469968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj, 139728975372992) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A, 139728976074912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'], 139733017758352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['second'].weight, 139722338354176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B, 139728976072368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['second'], 139733017757680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.base_layer, 139777663775312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout, 139728976081152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['second'], 139733018968944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj, 139728976188592) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A, 139728976184896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'], 139733021531232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['second'].weight, 139722338367136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B, 139728976190800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['second'], 139733018970384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.base_layer, 139777663775264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout, 139728976190416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['second'], 139733021798848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out, 139728975945472) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A, 139728975951952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'], 139729098437872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['second'].weight, 139722338365136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B, 139728975949888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B['second'], 139729098435760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.base_layer, 139777663775504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout, 139728975946672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['second'], 139729098437488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k, 139777663775648) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.weight, 139777664181776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q, 139777663775552) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.weight, 139777664181136) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.processor, 139777663774784) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1, 139777663774352) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm, 139777663774496) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu, 139777663774400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear, 139728975895840) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A, 139728975896704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['second'], 139726563997632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['second'].weight, 139722337014288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B, 139728975895888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B['second'], 139726563916192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.base_layer, 139777663774448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout, 139728975895600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout['second'], 139726563997728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2, 139777663775696) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context, 139777663776032) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net, 139777663776176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0], 139777663776128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj, 139728974653584) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A, 139728974644032) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'], 139733019320544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['second'].weight, 139722338357856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B, 139728974653632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['second'], 139733019325152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer, 139777663776224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout, 139728974651280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['second'], 139729096088144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1], 139777663776320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2], 139728974651664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A, 139728974657760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'], 139726562518320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['second'].weight, 139722338359536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B, 139728974647872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B['second'], 139726562518368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].base_layer, 139777663776368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout, 139728974645808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['second'], 139733019558816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context, 139777663774544) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm, 139777663774736) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu, 139777663774640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear, 139728975896224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A, 139728975898672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'], 139726563905728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['second'].weight, 139722337021008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B, 139728975898240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B['second'], 139726563905872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.base_layer, 139777663774688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout, 139728975896848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['second'], 139726563905824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context, 139777663775744) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11], 139777663774016) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff, 139777663777664) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net, 139777663777904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0], 139777663777856) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj, 139728974535152) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A, 139728974534960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'], 139729102445040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['second'].weight, 139722333981008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B, 139728974530784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['second'], 139729102454640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.base_layer, 139777663777952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout, 139728974536208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['second'], 139729102446912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1], 139777663778000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2], 139728974533472) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A, 139728974536640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['second'], 139729102440000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['second'].weight, 139722333981328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B, 139728974538512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B['second'], 139729102444752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].base_layer, 139777663778048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout, 139728974538080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout['second'], 139729102446720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn, 139777663776896) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k, 139728975482784) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A, 139728975482112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['second'], 139726564257904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['second'].weight, 139722333991968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B, 139728975483456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B['second'], 139726564247152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.base_layer, 139777663777040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout, 139728975482064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout['second'], 139726564257856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q, 139728977032192) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A, 139728977025616) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['second'], 139726563835232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['second'].weight, 139722333979568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B, 139728977026192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B['second'], 139726563824480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.base_layer, 139777663777136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout, 139728977032480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout['second'], 139726563835328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v, 139728975486240) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A, 139728975481776) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['second'], 139729102453728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['second'].weight, 139722333993568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B, 139728975484464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B['second'], 139729102451808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.base_layer, 139777663777232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout, 139728975485520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout['second'], 139729102448592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k, 139777663777088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.weight, 139777664182176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q, 139777663776992) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.weight, 139777664571232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out, 139777663777424) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0], 139728976310288) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A, 139728976308608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'], 139729102452480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['second'].weight, 139722333994288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B, 139728976310480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B['second'], 139729102448640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].base_layer, 139777663777472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout, 139728976301216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['second'], 139729102447056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1], 139777663777520) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj, 139728975480336) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A, 139728975495072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'], 139729102439904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['second'].weight, 139722333987008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B, 139728975494016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['second'], 139729102444080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.base_layer, 139777663777280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout, 139728975480528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['second'], 139729102444848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj, 139728977346128) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A, 139728977337728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'], 139729102445328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['second'].weight, 139722333993008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B, 139728977332448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['second'], 139729102453008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.base_layer, 139777663777376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout, 139728977344592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['second'], 139729102450800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj, 139728976133760) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A, 139728976122096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'], 139729102447104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['second'].weight, 139722333988128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B, 139728976122336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['second'], 139729102446768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.base_layer, 139777663777328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout, 139728976133616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['second'], 139729102454496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out, 139728974536352) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A, 139728974538176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'], 139729102451376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['second'].weight, 139722333978928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B, 139728974537120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B['second'], 139729102451280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.base_layer, 139777663777568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout, 139728974537264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['second'], 139729102449888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k, 139777663777712) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.weight, 139777664571952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q, 139777663777616) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.weight, 139777664178656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.processor, 139777663776848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1, 139777663776416) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm, 139777663776560) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu, 139777663776464) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear, 139728974655648) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A, 139728977035216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['second'], 139726564382160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['second'].weight, 139722338359136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B, 139728977021728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B['second'], 139726564382400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.base_layer, 139777663776512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout, 139728977022736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout['second'], 139726564392816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2, 139777663777760) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context, 139777663778096) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net, 139777663778240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0], 139777663778192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj, 139728974494400) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A, 139728974492960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'], 139726564160224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['second'].weight, 139722264459312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B, 139728974484848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['second'], 139726564149472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer, 139777663778288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout, 139728974490608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['second'], 139726564160320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1], 139777663778384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2], 139728976330320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A, 139728976322448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'], 139733012891584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['second'].weight, 139722264449472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B, 139728976318800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B['second'], 139733012892160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].base_layer, 139777663778432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout, 139728976322016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['second'], 139733018606800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context, 139777663776608) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm, 139777663776800) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu, 139777663776704) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear, 139728977020480) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A, 139728977028160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'], 139733022065328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['second'].weight, 139722333985168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B, 139728977034304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B['second'], 139733022073008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.base_layer, 139777663776752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout, 139728977020816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['second'], 139733012934720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context, 139777663777808) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12], 139777663776080) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff, 139777663615952) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net, 139777663616240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0], 139777663616192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj, 139728974844480) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A, 139728974849856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'], 139732772563616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['second'].weight, 139722264460112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B, 139728974850864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['second'], 139732772553056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.base_layer, 139777663616288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout, 139728974847936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['second'], 139732772554256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1], 139777663616336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2], 139728974741360) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A, 139728974736608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['second'], 139730414260944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['second'].weight, 139722264458592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B, 139728974739248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B['second'], 139730414262528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].base_layer, 139777663616384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout, 139728974738432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout['second'], 139732772556176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn, 139777663615184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k, 139728974553984) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A, 139728974556096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['second'], 139726563959680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['second'].weight, 139722264455472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B, 139728973206176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B['second'], 139726563960064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.base_layer, 139777663615328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout, 139728974559984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout['second'], 139733017891728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q, 139728974560128) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A, 139728974546784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['second'], 139726564517744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['second'].weight, 139722264453552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B, 139728974552976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B['second'], 139726564517888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.base_layer, 139777663615424) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout, 139728974551872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout['second'], 139726564517840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v, 139728973205312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A, 139728973207328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['second'], 139733020863904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['second'].weight, 139722264457952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B, 139728973205696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B['second'], 139733020864528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.base_layer, 139777663615520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout, 139728973214144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout['second'], 139726564029680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k, 139777663615376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.weight, 139777664568592) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q, 139777663615280) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.weight, 139777664569312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out, 139777663615712) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0], 139728974829920) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A, 139728974830064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'], 139732771481792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['second'].weight, 139722264459472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B, 139728974836784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B['second'], 139732771477616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].base_layer, 139777663615760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout, 139728974823536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['second'], 139732771479344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1], 139777663615808) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj, 139728973211984) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A, 139728973209824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'], 139726562604416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['second'].weight, 139722264451632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B, 139728973202960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['second'], 139733022666912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.base_layer, 139777663615568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout, 139728973216688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['second'], 139726562599232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj, 139728976184800) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A, 139728974830736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'], 139732771482704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['second'].weight, 139722264448992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B, 139728974832464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['second'], 139732771472912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.base_layer, 139777663615664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout, 139728974827232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['second'], 139726562680816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj, 139728973215776) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A, 139728973207664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'], 139733017947936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['second'].weight, 139722264452032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B, 139728973207520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['second'], 139733017949328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.base_layer, 139777663615616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout, 139728973202240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['second'], 139733022666816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out, 139728975500656) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A, 139728974855712) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'], 139732772567408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['second'].weight, 139722264446272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B, 139728974847168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B['second'], 139732772558336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.base_layer, 139777663615856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout, 139728975507472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['second'], 139732771477280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k, 139777663616000) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.weight, 139777660406176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q, 139777663615904) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.weight, 139777664572832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.processor, 139777663615136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1, 139777663778480) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm, 139777663778624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu, 139777663778528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear, 139728974328640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A, 139728974331376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['second'], 139733019714144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['second'].weight, 139722264452752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B, 139728974322304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B['second'], 139733019720624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.base_layer, 139777663778576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout, 139728974325616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout['second'], 139733020059696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2, 139777663616048) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context, 139777663616432) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net, 139777663616576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0], 139777663616528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj, 139728974740688) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A, 139728974741264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'], 139730414259696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['second'].weight, 139722264456912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B, 139728974737568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['second'], 139730414263344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer, 139777663616624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout, 139728974735936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['second'], 139730414270784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1], 139777663616720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2], 139728974729456) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A, 139728974726960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'], 139730414260992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['second'].weight, 139722264456512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B, 139728974728784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B['second'], 139730414268336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].base_layer, 139777663616768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout, 139728974726384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['second'], 139730414259744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context, 139777663778672) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm, 139777663615088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu, 139777663778768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear, 139728974461008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A, 139728974456928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'], 139726563862432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['second'].weight, 139722264455232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B, 139728974451696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B['second'], 139726563862384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.base_layer, 139777663615040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout, 139728974460768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['second'], 139733020411936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context, 139777663616096) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13], 139777663778144) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff, 139777663618064) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net, 139777663618304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0], 139777663618256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj, 139728973424832) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A, 139728973418592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'], 139730419017488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['second'].weight, 139722264655760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B, 139728973425600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['second'], 139730419014464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.base_layer, 139777663618352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout, 139728973426080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['second'], 139729102977632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1], 139777663618400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2], 139728973423056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A, 139728973419792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['second'], 139730419009904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['second'].weight, 139722264655280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B, 139728973423200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B['second'], 139730419016288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].base_layer, 139777663618448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout, 139728973418352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout['second'], 139730419010480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn, 139777663617296) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k, 139728974585456) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A, 139728974587664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['second'], 139729102973504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['second'].weight, 139722264447792) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B, 139728974591264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B['second'], 139729102970240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.base_layer, 139777663617440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout, 139728974592416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout['second'], 139729102969712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q, 139728974729072) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A, 139728974726192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['second'], 139729102972976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['second'].weight, 139722264459552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B, 139728974725280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B['second'], 139729102978736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.base_layer, 139777663617536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout, 139728974728400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout['second'], 139729102973360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v, 139728974579072) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A, 139728974582048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['second'], 139729102974272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['second'].weight, 139722264457392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B, 139728974591552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B['second'], 139729102975136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.base_layer, 139777663617632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout, 139728974592704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout['second'], 139729102974800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k, 139777663617488) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.weight, 139777659967648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q, 139777663617392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.weight, 139777659967728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out, 139777663617824) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0], 139728974382944) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A, 139728974390192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'], 139729102963808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['second'].weight, 139722264460752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B, 139728974391440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B['second'], 139729102963568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].base_layer, 139777663617872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout, 139728974381696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['second'], 139729102964240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1], 139777663617920) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj, 139728974593376) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A, 139728974578880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'], 139729102966592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['second'].weight, 139722264452272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B, 139728974587520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['second'], 139729102968368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.base_layer, 139777663617680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout, 139728974587280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['second'], 139729102964096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj, 139728974387312) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A, 139728974388752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'], 139729102966352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['second'].weight, 139722264460272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B, 139728974392160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['second'], 139729102968176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.base_layer, 139777663617776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout, 139728974394272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['second'], 139729102964864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj, 139728974384672) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A, 139728974382896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'], 139729102967504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['second'].weight, 139722264454272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B, 139728974381984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['second'], 139729102965824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.base_layer, 139777663617728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout, 139728974383376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['second'], 139729102967456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out, 139728974394800) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A, 139728974393072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'], 139729102975616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['second'].weight, 139722264461072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B, 139728974392400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B['second'], 139729102978256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.base_layer, 139777663617968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout, 139728974394368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['second'], 139729102976288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k, 139777663618112) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.weight, 139777664697824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q, 139777663618016) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.weight, 139777659967488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.processor, 139777663617248) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1, 139777663616816) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm, 139777663616960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu, 139777663616864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear, 139728974741024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A, 139728974738000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['second'], 139729102976576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['second'].weight, 139722264447312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B, 139728974729504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B['second'], 139729102977056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.base_layer, 139777663616912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout, 139728974741168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout['second'], 139729102979024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2, 139777663618160) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context, 139777663618496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net, 139777663618640) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0], 139777663618592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj, 139728973426896) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A, 139728973459904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'], 139730419010288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['second'].weight, 139722264644640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B, 139728973449824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['second'], 139730419017632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer, 139777663618688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout, 139728973428240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['second'], 139730419018016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1], 139777663618784) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2], 139728973462400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A, 139728973449008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'], 139732770120576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['second'].weight, 139722264647840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B, 139728973451744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B['second'], 139732770118272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].base_layer, 139777663618832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout, 139728973451936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['second'], 139732770122304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context, 139777663617008) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm, 139777663617200) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu, 139777663617104) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear, 139728974738816) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A, 139728974739152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'], 139729102977152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['second'].weight, 139722264449312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B, 139728974738624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B['second'], 139729102973696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.base_layer, 139777663617152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout, 139728974739392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['second'], 139729102975040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context, 139777663618208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14], 139777663616480) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff, 139777663620128) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net, 139777663620368) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0], 139777663620320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj, 139728975127808) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A, 139728975122960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'], 139729103796688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['second'].weight, 139722346551216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B, 139728975122672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['second'], 139729103797408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.base_layer, 139777663620416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout, 139728975128480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['second'], 139729103796928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1], 139777663620464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2], 139728975120320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A, 139728975119072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['second'], 139729103790112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['second'].weight, 139722346561136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B, 139728975119648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B['second'], 139729103784112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].base_layer, 139777663620512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout, 139728975120416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout['second'], 139729103787328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn, 139777663619360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k, 139728975048192) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A, 139728975050592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['second'], 139730414587424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['second'].weight, 139722347402704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B, 139728975043248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B['second'], 139730414587760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.base_layer, 139777663619504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout, 139728975040080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout['second'], 139730414594192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q, 139728975049152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A, 139728975048720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['second'], 139732773723184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['second'].weight, 139722264642320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B, 139728975049008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B['second'], 139732773720352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.base_layer, 139777663619600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout, 139728975049200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout['second'], 139732773725728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v, 139728975044208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A, 139728975041568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['second'], 139730414593904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['second'].weight, 139722625526304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B, 139728975045648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B['second'], 139730414592800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.base_layer, 139777663619696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout, 139728975046464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout['second'], 139730414592032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k, 139777663619552) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.weight, 139777659969328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q, 139777663619456) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.weight, 139777659969408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out, 139777663619888) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0], 139728974689280) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A, 139728974684864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'], 139729103793280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['second'].weight, 139722346554416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B, 139728974692304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B['second'], 139729103792032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].base_layer, 139777663619936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout, 139728974688176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['second'], 139729103793232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1], 139777663619984) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj, 139728975046944) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A, 139728975044736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'], 139729103788048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['second'].weight, 139722351055296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B, 139728973391920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['second'], 139729103788384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.base_layer, 139777663619744) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout, 139728975050352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['second'], 139722509320496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj, 139728974688416) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A, 139728974686784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'], 139729103788864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['second'].weight, 139722346558896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B, 139728974681456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['second'], 139729103793856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.base_layer, 139777663619840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout, 139728974686208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['second'], 139729103796400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj, 139728974691680) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A, 139728974687024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'], 139729103790880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['second'].weight, 139722346553376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B, 139728974692208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['second'], 139729103790448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.base_layer, 139777663619792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout, 139728974690864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['second'], 139729103791504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out, 139728974682032) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A, 139728975127520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'], 139729103794624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['second'].weight, 139722346560736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B, 139728975129584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B['second'], 139729103793664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.base_layer, 139777663620032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout, 139728974689712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['second'], 139729103795344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k, 139777663620176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.weight, 139777659969168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q, 139777663620080) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.weight, 139777659969248) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.processor, 139777663619312) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1, 139777663618880) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm, 139777663619024) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu, 139777663618928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear, 139728973451408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A, 139728973450208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['second'], 139732770126672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['second'].weight, 139722264644160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B, 139728973456064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B['second'], 139732770117600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.base_layer, 139777663618976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout, 139728973451024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout['second'], 139732770118848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2, 139777663620224) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context, 139777663620560) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net, 139777663620704) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0], 139777663620656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj, 139728973207376) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A, 139728976288144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'], 139729103783152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['second'].weight, 139722346545856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B, 139728976288096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['second'], 139729103785936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer, 139777663620752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout, 139728976282912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['second'], 139729103789344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1], 139777663620848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2], 139728976282816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A, 139728976282768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'], 139729103784304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['second'].weight, 139722346548496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B, 139728976290016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B['second'], 139729103791552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].base_layer, 139777663620896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout, 139728976281808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['second'], 139729103784256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context, 139777663619072) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm, 139777663619264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu, 139777663619168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear, 139728973459232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A, 139728973452560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'], 139730413510112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['second'].weight, 139722264641920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B, 139728973450736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B['second'], 139730413507520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.base_layer, 139777663619216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout, 139728973451072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['second'], 139730413517408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context, 139777663620272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15], 139777663618544) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff, 139777663622192) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net, 139777663622480) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0], 139777663622384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj, 139728974814496) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A, 139728974813296) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'], 139722485024608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['second'].weight, 139722341312256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B, 139728973566528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['second'], 139722485027680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.base_layer, 139777663622528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout, 139728974812768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['second'], 139722486473600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1], 139777663622576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2], 139728973577136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A, 139728973566384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['second'], 139722483433968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['second'].weight, 139722341307776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B, 139728973563360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B['second'], 139722483063136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].base_layer, 139777663622624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout, 139728973578000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout['second'], 139722482912704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn, 139777663621424) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k, 139728973901744) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A, 139728973905200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['second'], 139722505154864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['second'].weight, 139722341314656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B, 139728973892144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B['second'], 139722503678624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.base_layer, 139777663621568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout, 139728973904864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout['second'], 139722504675312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q, 139728973890080) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A, 139728973891376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['second'], 139722505889744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['second'].weight, 139722341311056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B, 139728973905728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B['second'], 139722505885328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.base_layer, 139777663621664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout, 139728973904192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout['second'], 139722506058816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v, 139728974970256) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A, 139728974970832) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['second'], 139722504432528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['second'].weight, 139722341303776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B, 139728974964256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B['second'], 139722502463088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.base_layer, 139777663621760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout, 139728974964304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout['second'], 139722506504032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k, 139777663621616) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.weight, 139777664583696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q, 139777663621520) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.weight, 139777664581776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out, 139777663621952) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0], 139728973325616) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A, 139728973322592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'], 139722489023792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['second'].weight, 139722341316736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B, 139728973329792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B['second'], 139722487788592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].base_layer, 139777663622000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout, 139728973325712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['second'], 139722491836352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1], 139777663622048) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj, 139728974273008) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A, 139728974278576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'], 139722501661232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['second'].weight, 139722341317616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B, 139728974279008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['second'], 139722501656624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.base_layer, 139777663621808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout, 139728974279536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['second'], 139722501251200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj, 139728973329936) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A, 139728973328928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'], 139722494761360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['second'].weight, 139722341316496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B, 139728973331136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['second'], 139722494766304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.base_layer, 139777663621904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout, 139728973329120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['second'], 139722493276368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj, 139728975132944) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A, 139728973318560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'], 139722498179664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['second'].weight, 139722341316176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B, 139728973319472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['second'], 139722497486624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.base_layer, 139777663621856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout, 139728975180704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['second'], 139722501921984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out, 139728973316352) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A, 139728973317120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'], 139722486923120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['second'].weight, 139722341317296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B, 139728973316496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B['second'], 139722488297984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.base_layer, 139777663622096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout, 139728973324752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['second'], 139722488782688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k, 139777663622240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.weight, 139777664574432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q, 139777663622144) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.weight, 139777664746656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.processor, 139777663621376) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1, 139777663620944) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm, 139777663621088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu, 139777663620992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear, 139728973306544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A, 139728973309520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['second'], 139729103786368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['second'].weight, 139722346555536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B, 139728973309088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B['second'], 139729103783200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.base_layer, 139777663621040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout, 139728973299968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout['second'], 139729103785264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2, 139777663622288) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context, 139777663622672) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net, 139777663622816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0], 139777663622768) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj, 139728973570896) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A, 139728973562208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'], 139722479559312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['second'].weight, 139722341307456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B, 139728973570128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['second'], 139722480082064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer, 139777663622864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout, 139728973570944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['second'], 139722482033056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1], 139777663622960) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2], 139728973565616) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A, 139728973576128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'], 139722479966656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['second'].weight, 139722341308816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B, 139728973564272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B['second'], 139722479961232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].base_layer, 139777663623008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout, 139728973565808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['second'], 139722480183104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context, 139777663621136) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm, 139777663621328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu, 139777663621232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear, 139728973890800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A, 139728973905488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'], 139722509077952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['second'].weight, 139722341303856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B, 139728973890128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B['second'], 139722506942208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.base_layer, 139777663621280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout, 139728973905344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['second'], 139722508083280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context, 139777663622336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16], 139777663620608) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff, 139777663624352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net, 139777663624592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0], 139777663624544) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj, 139728972213568) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A, 139728974061024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'], 139722379117568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['second'].weight, 139722342209536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B, 139728974066448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['second'], 139722379126496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.base_layer, 139777663624640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout, 139728974068176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['second'], 139722383751344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1], 139777663624688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2], 139728974278192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A, 139728973822512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['second'], 139722377650496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['second'].weight, 139722342209936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B, 139728973823472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B['second'], 139722377646512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].base_layer, 139777663624736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout, 139728973810224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout['second'], 139722379570496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn, 139777663623584) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k, 139728973512464) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A, 139728973502192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['second'], 139722469652752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['second'].weight, 139722341307696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B, 139728973508960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B['second'], 139722469129520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.base_layer, 139777663623728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout, 139728973512224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout['second'], 139722471310720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q, 139728973510448) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A, 139728973499216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['second'], 139722471689184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['second'].weight, 139722341310576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B, 139728973500080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B['second'], 139722471182192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.base_layer, 139777663623824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout, 139728973500656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout['second'], 139722472133568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v, 139728972211408) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A, 139728972217504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['second'], 139722467688208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['second'].weight, 139722342215616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B, 139728972203872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B['second'], 139722467675248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.base_layer, 139777663623920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout, 139728972212512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout['second'], 139722468802128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k, 139777663623776) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.weight, 139777664327232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q, 139777663623680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.weight, 139777664594016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out, 139777663624112) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0], 139728972202096) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A, 139728972204352) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'], 139722387328544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['second'].weight, 139722342214896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B, 139728972203584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B['second'], 139722387327872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].base_layer, 139777663624160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout, 139728972204832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['second'], 139722388248784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1], 139777663624208) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj, 139728972209584) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A, 139728972204784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'], 139722395699520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['second'].weight, 139722342211616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B, 139728972205504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['second'], 139722395705616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.base_layer, 139777663623968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout, 139728972217264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['second'], 139722399564176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj, 139728972215824) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A, 139728972215200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'], 139722391069088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['second'].weight, 139722342214496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B, 139728972214912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['second'], 139722472492240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.base_layer, 139777663624064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout, 139728972216352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['second'], 139722469508048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj, 139728972216640) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A, 139728972218176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'], 139722470812416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['second'].weight, 139722342212096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B, 139728972217648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['second'], 139722393687168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.base_layer, 139777663624016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout, 139728972209872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['second'], 139722393509392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out, 139728972211792) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A, 139728972213952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'], 139722383884192) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['second'].weight, 139722342209056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B, 139728972212464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B['second'], 139722385723440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.base_layer, 139777663624256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout, 139728972216400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['second'], 139722384272080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k, 139777663624400) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.weight, 139777664592896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q, 139777663624304) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.weight, 139777664595776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.processor, 139777663623536) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1, 139777663623056) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm, 139777663623248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu, 139777663623152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear, 139728973566000) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A, 139728973569504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['second'], 139722476549168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['second'].weight, 139722341317696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B, 139728973563840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B['second'], 139722476545376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.base_layer, 139777663623200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout, 139728973569216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout['second'], 139722475912064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2, 139777663624448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context, 139777663624784) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net, 139777663624928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0], 139777663624880) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj, 139728975241872) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A, 139728975233280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'], 139722368366288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['second'].weight, 139722342205456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B, 139728975249216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['second'], 139722371923248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer, 139777663624976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout, 139728975246912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['second'], 139722371166176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1], 139777663625072) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2], 139728972455344) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A, 139728972463312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'], 139722362254576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['second'].weight, 139722342206096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B, 139728972452704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B['second'], 139722364828400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].base_layer, 139777663625120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout, 139728972451552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['second'], 139722362379264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context, 139777663623296) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm, 139777663623488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu, 139777663623392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear, 139728973578096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A, 139728973506272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'], 139722472357840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['second'].weight, 139722341315856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B, 139728973509152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B['second'], 139722476041456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.base_layer, 139777663623440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout, 139728973563456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['second'], 139722473865664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context, 139777663624496) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17], 139777663622720) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff, 139777663626416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net, 139777663626656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0], 139777663626608) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj, 139728974249120) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A, 139728974247968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'], 139722344943280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['second'].weight, 139722342215936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B, 139728974246144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['second'], 139722344946688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.base_layer, 139777663626704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout, 139728974248640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['second'], 139722344944768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1], 139777663626752) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2], 139728974246912) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A, 139728974246864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['second'], 139722344944384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['second'].weight, 139722342123056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B, 139728974245952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B['second'], 139722344941120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].base_layer, 139777663626800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout, 139728974247008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout['second'], 139722344942704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn, 139777663625648) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k, 139728972369440) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A, 139728972366224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['second'], 139722366726688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['second'].weight, 139722342215776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B, 139728972380480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B['second'], 139722366718432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.base_layer, 139777663625792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout, 139728972380288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout['second'], 139722363579184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q, 139728972371600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A, 139728972371024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['second'], 139722363670144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['second'].weight, 139722342209216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B, 139728972375680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B['second'], 139722363676048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.base_layer, 139777663625888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout, 139728972378896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout['second'], 139722348546224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v, 139728973750064) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A, 139728973742720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['second'], 139722344946496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['second'].weight, 139722342217216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B, 139728973750208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B['second'], 139722344947696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.base_layer, 139777663625984) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout, 139728973749632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout['second'], 139722344946880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k, 139777663625840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.weight, 139777675123168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q, 139777663625744) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.weight, 139777660095440) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out, 139777663626176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0], 139728974236016) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A, 139728974237456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'], 139722344947360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['second'].weight, 139722342218176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B, 139728974236976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B['second'], 139722344948608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].base_layer, 139777663626224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout, 139728974236256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['second'], 139722344944144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1], 139777663626272) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj, 139728975195984) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A, 139728975194880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'], 139722344946256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['second'].weight, 139722342217136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B, 139728975195696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['second'], 139722344944528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.base_layer, 139777663626032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout, 139728975192624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['second'], 139722344943760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj, 139728973716768) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A, 139728974235824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'], 139722344943856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['second'].weight, 139722342217776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B, 139728974236880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['second'], 139722344944720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.base_layer, 139777663626128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout, 139728974236496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['second'], 139722344943808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj, 139728973722000) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A, 139728973722240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'], 139722344940832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['second'].weight, 139722342217456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B, 139728973714944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['second'], 139722344941696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.base_layer, 139777663626080) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout, 139728973721904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['second'], 139722344940208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out, 139728974237408) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A, 139728974247680) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'], 139722344941648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['second'].weight, 139722342218576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B, 139728974244848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B['second'], 139722344948320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.base_layer, 139777663626320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout, 139728974242208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['second'], 139722344943040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k, 139777663626464) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.weight, 139777664329392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q, 139777663626368) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.weight, 139777664585456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.processor, 139777663625600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1, 139777663625168) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm, 139777663625312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu, 139777663625216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear, 139728972458800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A, 139728972463744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['second'], 139722343060656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['second'].weight, 139722342206656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B, 139728972374240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B['second'], 139722345110336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.base_layer, 139777663625264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout, 139728972453472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout['second'], 139722355666480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2, 139777663626512) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context, 139777663626848) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net, 139777663626992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0], 139777663626944) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj, 139728974234960) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A, 139728974237216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'], 139722344946640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['second'].weight, 139722342135536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B, 139728974235584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['second'], 139722344948992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer, 139777663627040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout, 139728974235680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['second'], 139722344941216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1], 139777663627136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2], 139728974062464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A, 139728975001632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'], 139722344947072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['second'].weight, 139722342129296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B, 139728974994192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B['second'], 139722344940112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].base_layer, 139777663627184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout, 139728975002064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['second'], 139722344947408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context, 139777663625360) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm, 139777663625552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu, 139777663625456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear, 139728972375728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A, 139728972376304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'], 139722343297952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['second'].weight, 139722342204096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B, 139728972372272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B['second'], 139722348544352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.base_layer, 139777663625504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout, 139728972371792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['second'], 139722343234096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context, 139777663626560) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18], 139777663624832) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff, 139777663628480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net, 139777663628720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0], 139777663628672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj, 139728972775184) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A, 139728972762416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'], 139722264020832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['second'].weight, 139722341225136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B, 139728972762320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['second'], 139722264031248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.base_layer, 139777663628768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout, 139728972760784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['second'], 139722264032304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1], 139777663628816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2], 139728974236064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A, 139728972688512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['second'], 139722264052304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['second'].weight, 139722341225696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B, 139728972691776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B['second'], 139722264059696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].base_layer, 139777663628864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout, 139728972682800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout['second'], 139722264019344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn, 139777663627712) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k, 139728973977088) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A, 139728973976896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['second'], 139728988962720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['second'].weight, 139722341232336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B, 139728973975408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B['second'], 139722264026736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.base_layer, 139777663627856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout, 139728973977136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout['second'], 139722343345424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q, 139728973973920) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A, 139728973979296) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['second'], 139728989246304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['second'].weight, 139722342123376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B, 139728973976464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B['second'], 139728989471312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.base_layer, 139777663627952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout, 139728973972816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout['second'], 139722344944432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v, 139728974183152) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A, 139728974168848) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['second'], 139722264029376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['second'].weight, 139722341232496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B, 139728974184256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B['second'], 139722264032208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.base_layer, 139777663628048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout, 139728974182144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout['second'], 139722264034176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k, 139777663627904) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.weight, 139777664592016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q, 139777663627808) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.weight, 139777664592096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out, 139777663628240) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0], 139728972766976) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A, 139728972761744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'], 139722264035040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['second'].weight, 139722341224416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B, 139728972767168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B['second'], 139722264030336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].base_layer, 139777663628288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout, 139728972774128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['second'], 139722264025056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_out[0].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_out[0].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_out[0].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1], 139777663628336) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj, 139728974183296) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A, 139728974182672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'], 139722264030432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['second'].weight, 139722341234096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B, 139728974169088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['second'], 139722264021840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.base_layer, 139777663628096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout, 139728974181616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['second'], 139722264031632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_k_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_k_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj, 139728972766208) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A, 139728972767552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'], 139722264022032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['second'].weight, 139722341224096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B, 139728972767600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['second'], 139722264029904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.base_layer, 139777663628192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout, 139728972766592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['second'], 139722264024048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_q_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_q_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj, 139728974173072) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A, 139728974173504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'], 139722264030288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['second'].weight, 139722341221456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B, 139728972774080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['second'], 139722264025008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.base_layer, 139777663628144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout, 139728974178928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['second'], 139722264034608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_v_proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_v_proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out, 139728972767888) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A, 139728972760208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'], 139722264019584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['second'].weight, 139722341224816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B, 139728972768608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B['second'], 139722264019824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.base_layer, 139777663628384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout, 139728972760544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['second'], 139722264021984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_add_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_add_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_add_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k, 139777663628528) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.weight, 139777664758736) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q, 139777663628432) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.weight, 139777664333392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.processor, 139777663627664) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1, 139777663627232) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm, 139777663627376) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu, 139777663627280) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear, 139728974995488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A, 139728973972048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['second'], 139722344946400) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['second'].weight, 139722342129056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B, 139728973971904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B['second'], 139722344944240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.base_layer, 139777663627328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout, 139728973978864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout['second'], 139722344945920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].norm1.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2, 139777663628576) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context, 139777663628912) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net, 139777663629056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0], 139777663629008) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj, 139728972759776) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A, 139728972501184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'], 139722264066656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['second'].weight, 139722341226096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B, 139728972510736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['second'], 139722264059552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer, 139777663629104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout, 139728972512896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['second'], 139722264059600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1], 139777663629200) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2], 139728972498400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A, 139728972505792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'], 139722264058304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['second'].weight, 139722341226576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B, 139728972500464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B['second'], 139722264052016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].base_layer, 139777663629248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout, 139728972510160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['second'], 139722264051920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[2].scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[2].scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[2].use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context, 139777663627424) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm, 139777663627616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu, 139777663627520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear, 139728973986400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A, 139728973986208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'], 139722344940544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['second'].weight, 139722342130576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B, 139728973986160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B['second'], 139722344949904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.base_layer, 139777663627568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout, 139728973986688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['second'], 139722344939728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1_context.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].norm1_context.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1_context.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context, 139777663628624) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=L['self'].single_transformer_blocks, accessed_by=DictGetItemGuardAccessor(single_transformer_blocks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks, 139777663628960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0], 139777663626896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn, 139777663629776) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k, 139728973631584) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A, 139728973630048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'], 139722264060704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['second'].weight, 139722341230416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B, 139728973628368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B['second'], 139722264060656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.base_layer, 139777663629920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout, 139728973631632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['second'], 139722264061664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q, 139728975203488) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A, 139728975210112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'], 139722264054800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['second'].weight, 139722341229136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B, 139728975214384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B['second'], 139722264054560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.base_layer, 139777663630016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout, 139728975203872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['second'], 139722264057296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v, 139728973627696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A, 139728973629568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'], 139722264057680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['second'].weight, 139722341229696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B, 139728973629664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B['second'], 139722264052208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.base_layer, 139777663630064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout, 139728973628176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['second'], 139722264061616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k, 139777663629968) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.weight, 139777659813168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q, 139777663629824) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.weight, 139777660410496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.processor, 139777663629728) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm, 139777663629344) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm, 139777663629488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu, 139777663629392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear, 139728972497728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A, 139728972498064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'], 139722264057152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['second'].weight, 139722341221216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B, 139728972498736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B['second'], 139722264064784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.base_layer, 139777663629440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout, 139728972497248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['second'], 139722264061424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp, 139777663629632) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp, 139728972501952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A, 139728972504400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'], 139722264064496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['second'].weight, 139722341222016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B, 139728972512464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B['second'], 139722264056144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.base_layer, 139777663629584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout, 139728972510640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['second'], 139722264056960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out, 139728975204064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A, 139728975202480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['second'], 139722264059312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['second'].weight, 139722341229296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B, 139728975204304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B['second'], 139722264062480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.base_layer, 139777663629680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout, 139728975202432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout['second'], 139722264059264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1], 139777663629296) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn, 139777663630544) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k, 139728974145968) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A, 139728974141264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'], 139722264065312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['second'].weight, 139722341236176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B, 139728974093744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B['second'], 139722264063680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.base_layer, 139777663630688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout, 139728974147696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['second'], 139722264065264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q, 139728973400240) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A, 139728973410608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'], 139722264067088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['second'].weight, 139722341235776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B, 139728974137520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B['second'], 139722264064592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.base_layer, 139777663630784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout, 139728973400336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['second'], 139722264057632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v, 139728974100320) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A, 139728974092016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'], 139722264067280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['second'].weight, 139722341234816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B, 139728974102240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B['second'], 139722264067136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.base_layer, 139777663630832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout, 139728974100608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['second'], 139722264067232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k, 139777663630736) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.weight, 139777660121344) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q, 139777663630592) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.weight, 139777664173856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.processor, 139777663630496) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm, 139777663630160) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm, 139777663630304) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu, 139777663630208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear, 139728973632112) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A, 139728973630864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'], 139722264058784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['second'].weight, 139722341233456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B, 139728973632160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B['second'], 139722264058448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.base_layer, 139777663630256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout, 139728973643392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['second'], 139722264058544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp, 139777663630400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp, 139728973863936) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A, 139728973866144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'], 139722264058160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['second'].weight, 139722341235536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B, 139728973863024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B['second'], 139722264066032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.base_layer, 139777663630352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout, 139728973859856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['second'], 139722264058208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out, 139728973410368) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A, 139728973401008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['second'], 139722264062960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['second'].weight, 139722341235456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B, 139728973399616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B['second'], 139722264063008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.base_layer, 139777663630448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout, 139728973406768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout['second'], 139722264063104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2], 139777663630112) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn, 139777663631312) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k, 139728972985248) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A, 139728972986688) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'], 139722248314000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['second'].weight, 139722346211632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B, 139728972986976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B['second'], 139722248313856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.base_layer, 139777662337184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout, 139728972979296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['second'], 139722248313952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q, 139728972288048) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A, 139728972295200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'], 139722248314384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['second'].weight, 139722346210912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B, 139728972290448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B['second'], 139722248314528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.base_layer, 139777662337280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout, 139728972292320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['second'], 139722248314432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v, 139728972974880) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A, 139728972976896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'], 139722248312896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['second'].weight, 139722346210272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B, 139728972984528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B['second'], 139722248313376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.base_layer, 139777662337328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout, 139728972986352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['second'], 139722248312848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k, 139777662337232) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.weight, 139777660414656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q, 139777662337088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.weight, 139777664172096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.processor, 139777663631264) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm, 139777663630928) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm, 139777663631072) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu, 139777663630976) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear, 139728972338208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A, 139728972730464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'], 139722264057584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['second'].weight, 139722367109216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B, 139728972742080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B['second'], 139722264060320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.base_layer, 139777663631024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout, 139728972335136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['second'], 139722264055184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp, 139777663631168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp, 139728972727200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A, 139728972296112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'], 139722264059936) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['second'].weight, 139722367116416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B, 139728972287280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B['second'], 139722264055856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.base_layer, 139777663631120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout, 139728972286176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['second'], 139722264054944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out, 139728972288768) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A, 139728972287760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['second'], 139722264057536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['second'].weight, 139722366926768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B, 139728972288816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B['second'], 139722248314912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.base_layer, 139777663631216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout, 139728972288384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout['second'], 139722264057440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3], 139777663630880) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn, 139777662337808) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k, 139728973077600) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A, 139728973136992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'], 139722248321920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['second'].weight, 139722346213392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B, 139728973145872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B['second'], 139722248315968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.base_layer, 139777662337952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout, 139728973080816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['second'], 139722248315440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q, 139728973080432) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A, 139728973086624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'], 139722248316256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['second'].weight, 139722346204192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B, 139728973082400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B['second'], 139722248318896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.base_layer, 139777662338048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout, 139728973079136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['second'], 139722248307808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v, 139728973140496) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A, 139728973149616) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'], 139722248319280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['second'].weight, 139722346202992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B, 139728973151872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B['second'], 139722248319376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.base_layer, 139777662338096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout, 139728973139872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['second'], 139722248321344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k, 139777662338000) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.weight, 139777660434480) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q, 139777662337856) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.weight, 139777664183776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.processor, 139777662337760) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm, 139777662337424) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm, 139777662337568) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu, 139777662337472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear, 139728972975984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A, 139728972979200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'], 139722248312800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['second'].weight, 139722346202752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B, 139728972985968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B['second'], 139722248313184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.base_layer, 139777662337520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout, 139728972983280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['second'], 139722248313424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp, 139777662337664) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp, 139728973081968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A, 139728973081488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'], 139722248312752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['second'].weight, 139722346212032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B, 139728973082160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B['second'], 139722248312608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.base_layer, 139777662337616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout, 139728973073184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['second'], 139722248312704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out, 139728973078032) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A, 139728973077648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['second'], 139722248320960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['second'].weight, 139722346212832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B, 139728973078800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B['second'], 139722248321440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.base_layer, 139777662337712) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout, 139728973078992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout['second'], 139722248319616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4], 139777662337376) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn, 139777662338576) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k, 139728972269024) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A, 139728972272432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'], 139722248318704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['second'].weight, 139722369580256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B, 139728972271184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B['second'], 139722248311792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.base_layer, 139777662338720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout, 139728972277232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['second'], 139722248319232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q, 139728974045168) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A, 139728974038208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'], 139722248315680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['second'].weight, 139722346203552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B, 139728974042096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B['second'], 139722248322208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.base_layer, 139777662338816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout, 139728974048432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['second'], 139722248311072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v, 139728972275936) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A, 139728972269072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'], 139722248306752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['second'].weight, 139722625503360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B, 139728972269648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B['second'], 139722248315584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.base_layer, 139777662338864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout, 139728972276752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['second'], 139722248310208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k, 139777662338768) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.weight, 139777664176976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q, 139777662338624) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.weight, 139777660379648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.processor, 139777662338528) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm, 139777662338192) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm, 139777662338336) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu, 139777662338240) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear, 139728973143520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A, 139728973140736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'], 139722248320768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['second'].weight, 139722346207632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B, 139728973144384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B['second'], 139722248320816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.base_layer, 139777662338288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout, 139728973145680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['second'], 139722248319568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp, 139777662338432) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp, 139728973147696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A, 139728973147024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'], 139722248315824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['second'].weight, 139722346203472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B, 139728973144144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B['second'], 139722248308432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.base_layer, 139777662338384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout, 139728973138240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['second'], 139722248320096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out, 139728973149664) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A, 139728973146016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['second'], 139722248307712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['second'].weight, 139722346211152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B, 139728973146832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B['second'], 139722248317792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.base_layer, 139777662338480) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout, 139728973152160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout['second'], 139722248318752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5], 139777662338144) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn, 139777662339344) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k, 139728972551920) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A, 139728972551152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'], 139722344830656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['second'].weight, 139722620008016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B, 139728972553648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B['second'], 139722344839536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.base_layer, 139777662339488) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout, 139728972552592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['second'], 139722344833392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q, 139728973662624) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A, 139728972557104) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'], 139722344838000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['second'].weight, 139722620010256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B, 139728972553936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B['second'], 139722344835456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.base_layer, 139777662339584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout, 139728972556336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['second'], 139722344839920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v, 139728972554992) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A, 139728972816272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'], 139722344830224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['second'].weight, 139722625545488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B, 139728972813872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B['second'], 139722344831280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.base_layer, 139777662339632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout, 139728972558256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['second'], 139722344830368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k, 139777662339536) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.weight, 139781198839008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q, 139777662339392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.weight, 139777664579472) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.processor, 139777662339296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm, 139777662338960) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm, 139777662339104) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu, 139777662339008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear, 139728972282128) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A, 139728972268880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'], 139722248321392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['second'].weight, 139722617180128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B, 139728972270752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B['second'], 139722248320528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.base_layer, 139777662339056) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout, 139728972277568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['second'], 139722248319328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp, 139777662339200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp, 139728972271328) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A, 139728972269888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'], 139722344840832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['second'].weight, 139722617180608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B, 139728972276704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B['second'], 139722344828832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.base_layer, 139777662339152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout, 139728972272144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['second'], 139722248309056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out, 139728973673328) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A, 139728973661136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['second'], 139722344838768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['second'].weight, 139722625042128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B, 139728973674384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B['second'], 139722344838576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.base_layer, 139777662339248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout, 139728973676064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout['second'], 139722344839344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6], 139777662338912) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn, 139777662340112) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k, 139728971307072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A, 139728971302176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'], 139722475017200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['second'].weight, 139722625533248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B, 139728971307168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B['second'], 139722474170464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.base_layer, 139777662340256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout, 139728971310960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['second'], 139722476990144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q, 139728971315280) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A, 139728971314416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'], 139722491498256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['second'].weight, 139722625545088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B, 139728971308080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B['second'], 139722491499072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.base_layer, 139777662340352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout, 139728971312352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['second'], 139722492617792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v, 139728971204736) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A, 139728971213184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'], 139722474165520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['second'].weight, 139722625534048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B, 139728971212224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B['second'], 139722474166048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.base_layer, 139777662340400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout, 139728971209968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['second'], 139722475281456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k, 139777662340304) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.weight, 139777664186496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q, 139777662340160) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.weight, 139777660410336) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.processor, 139777662340064) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm, 139777662339728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm, 139777662339872) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu, 139777662339776) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear, 139728972812576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A, 139728972811568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'], 139729092548960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['second'].weight, 139722625532528) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B, 139728972811376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B['second'], 139722490393072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.base_layer, 139777662339824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout, 139728972813008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['second'], 139722507219920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp, 139777662339968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp, 139728972818288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A, 139728972818000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'], 139722495931776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['second'].weight, 139722625548128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B, 139728972623360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B['second'], 139722495276320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.base_layer, 139777662339920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout, 139728972818192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['second'], 139722494642928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out, 139728972618224) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A, 139728971314128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['second'], 139722495277568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['second'].weight, 139722625537808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B, 139728971310624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B['second'], 139722495273872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.base_layer, 139777662340016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout, 139728971302416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout['second'], 139722493509792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7], 139777662339680) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn, 139777662340880) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k, 139728971383280) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A, 139728971281456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'], 139722476784896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['second'].weight, 139722618743488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B, 139728971273584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B['second'], 139722378875888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.base_layer, 139777662341024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout, 139728971387408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['second'], 139722494091632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q, 139728971398544) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A, 139728971396816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'], 139722506329424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['second'].weight, 139722625544448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B, 139728971397728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B['second'], 139722381321232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.base_layer, 139777662341120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout, 139728971397920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['second'], 139722473718784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v, 139728971281984) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A, 139728971282608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'], 139722367388304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['second'].weight, 139722618732848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B, 139728971283136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B['second'], 139722367393200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.base_layer, 139777662341168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout, 139728971279536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['second'], 139722378878672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k, 139777662341072) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.weight, 139777664181696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q, 139777662340928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.weight, 139777664579392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.processor, 139777662340832) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm, 139777662340496) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm, 139777662340640) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu, 139777662340544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear, 139728971250560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A, 139728971235728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'], 139722477241872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['second'].weight, 139722625546288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B, 139728971250752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B['second'], 139722475758848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.base_layer, 139777662340592) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout, 139728971239376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['second'], 139722476468880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp, 139777662340736) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp, 139728971392352) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A, 139728971393696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'], 139722396638400) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['second'].weight, 139722625536048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B, 139728971393168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B['second'], 139722396644064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.base_layer, 139777662340688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout, 139728971392208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['second'], 139722475763504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out, 139728971391584) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A, 139728971390528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['second'], 139722396219472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['second'].weight, 139722625536368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B, 139728971397968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B['second'], 139722396213568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.base_layer, 139777662340784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout, 139728971391920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout['second'], 139722468710464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8], 139777662340448) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn, 139777662341648) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k, 139728972261088) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A, 139728972936480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'], 139722344788128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['second'].weight, 139722263847984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B, 139728972936048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B['second'], 139722344779296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.base_layer, 139777662341792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout, 139728972926688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['second'], 139722344783904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q, 139728972251296) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A, 139728972259696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'], 139722344786496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['second'].weight, 139722618742768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B, 139728972260464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B['second'], 139722344788560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.base_layer, 139777662341888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout, 139728972251440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['second'], 139722344778096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v, 139728972929712) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A, 139728972930288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'], 139722372199280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['second'].weight, 139722618748048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B, 139728972936864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B['second'], 139722368477040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.base_layer, 139777662341936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout, 139728972938352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['second'], 139722372194816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k, 139777662341840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.weight, 139777664477008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q, 139777662341696) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.weight, 139777664475008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.processor, 139777662341600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm, 139777662341264) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm, 139777662341408) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu, 139777662341312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear, 139728972659152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A, 139728972660640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'], 139722372574768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['second'].weight, 139722618732768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B, 139728972659584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B['second'], 139722372569200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.base_layer, 139777662341360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout, 139728972644464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['second'], 139728971272432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp, 139777662341504) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp, 139728972652096) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A, 139728972640800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'], 139722372920176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['second'].weight, 139722618739648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B, 139728972632928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B['second'], 139728973900352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.base_layer, 139777662341456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout, 139728972646048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['second'], 139722369236512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out, 139728972631008) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A, 139728972632160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['second'], 139722344781264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['second'].weight, 139722618745328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B, 139728972642336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B['second'], 139722344782368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.base_layer, 139777662341552) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout, 139728972643728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout['second'], 139722344784192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9], 139777662341216) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn, 139777662342416) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k, 139728971115264) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A, 139728971113536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'], 139722347657296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['second'].weight, 139722625283264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B, 139728971106624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B['second'], 139722347657968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.base_layer, 139777662342560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout, 139728971117088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['second'], 139722347657824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q, 139728971119152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A, 139728971115552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'], 139722347657872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['second'].weight, 139722618745008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B, 139728971109888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B['second'], 139722347658448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.base_layer, 139777662342656) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout, 139728971118672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['second'], 139722347658256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v, 139728971106096) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A, 139728971106528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'], 139722347656432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['second'].weight, 139722625277584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B, 139728971119248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B['second'], 139722347657152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.base_layer, 139777662342704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout, 139728971105856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['second'], 139722347656192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k, 139777662342608) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.weight, 139777664476128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q, 139777662342464) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.weight, 139777664479568) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.processor, 139777662342368) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm, 139777662342032) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm, 139777662342176) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu, 139777662342080) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear, 139728972826368) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A, 139728972833904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'], 139722364192176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['second'].weight, 139722618736048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B, 139728972829728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B['second'], 139722364200864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.base_layer, 139777662342128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout, 139728972836016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['second'], 139722369129040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp, 139777662342272) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp, 139728971175904) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A, 139728971181712) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'], 139722364924976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['second'].weight, 139722618746128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B, 139728971185792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B['second'], 139722347648512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.base_layer, 139777662342224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout, 139728971184592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['second'], 139722363768928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out, 139728971105280) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A, 139728971117856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['second'], 139722347658976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['second'].weight, 139722618739808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B, 139728971120064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B['second'], 139722347659072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.base_layer, 139777662342320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout, 139728971119104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout['second'], 139722347659168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10], 139777662341984) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn, 139777662343184) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k, 139728971470960) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A, 139728971468704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'], 139722347652976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['second'].weight, 139722625285184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B, 139728971470816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B['second'], 139722347653600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.base_layer, 139777662343328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout, 139728971471008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['second'], 139722347653216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q, 139728971466352) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A, 139728971471488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'], 139722347653888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['second'].weight, 139722625283184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B, 139728971468944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B['second'], 139722347653984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.base_layer, 139777662343424) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout, 139728971466064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['second'], 139722347653504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v, 139728971472544) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A, 139728971473216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'], 139722347652544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['second'].weight, 139722625279904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B, 139728971473024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B['second'], 139722347651968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.base_layer, 139777662343472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout, 139728971472448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['second'], 139722347652496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k, 139777662343376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.weight, 139777659800128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q, 139777662343232) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.weight, 139777664329632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.processor, 139777662343136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm, 139777662342800) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm, 139777662342944) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu, 139777662342848) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear, 139728971114256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A, 139728971117040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'], 139722347655952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['second'].weight, 139722625284304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B, 139728971116176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B['second'], 139722347656480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.base_layer, 139777662342896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout, 139728971114736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['second'], 139722347656288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp, 139777662343040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp, 139728971106480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A, 139728971106288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'], 139722347655760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['second'].weight, 139722625269904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B, 139728971105808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B['second'], 139722347655472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.base_layer, 139777662342992) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout, 139728971106768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['second'], 139722347655712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out, 139728971469568) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A, 139728971469808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['second'], 139722347654176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['second'].weight, 139722625270864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B, 139728971470096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B['second'], 139722347654944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.base_layer, 139777662343088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout, 139728971469520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout['second'], 139722347655040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11], 139777662342752) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn, 139777662343952) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k, 139728972035072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A, 139728972026192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'], 139722347648608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['second'].weight, 139722625277024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B, 139728972030848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B['second'], 139722347648464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.base_layer, 139777662344096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout, 139728972031328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['second'], 139722347648704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q, 139728971480368) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A, 139728971480560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'], 139722347650672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['second'].weight, 139722625277104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B, 139728971480704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B['second'], 139722347650240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.base_layer, 139777662344192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout, 139728971480272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['second'], 139722347650720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v, 139728972022112) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A, 139728972022304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'], 139722347648128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['second'].weight, 139722625275904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B, 139728971155296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B['second'], 139722347643616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.base_layer, 139777662344240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout, 139728972022016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['second'], 139722347647984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k, 139777662344144) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.weight, 139777660418096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q, 139777662344000) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.weight, 139777659854400) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.processor, 139777662343904) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm, 139777662343568) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm, 139777662343712) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu, 139777662343616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear, 139728971475952) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A, 139728971476624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'], 139722347651200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['second'].weight, 139722625285504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B, 139728971477008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B['second'], 139722347651824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.base_layer, 139777662343664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout, 139728971473888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['second'], 139722347651920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp, 139777662343808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp, 139728971479600) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A, 139728971478256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'], 139722347643328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['second'].weight, 139722625280384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B, 139728971479168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B['second'], 139722347650960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.base_layer, 139777662343760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout, 139728971479696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['second'], 139722347644864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out, 139728971464816) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A, 139728971478784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['second'], 139722347643232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['second'].weight, 139722625278064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B, 139728971478640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B['second'], 139722347643520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.base_layer, 139777662343856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout, 139728971477344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout['second'], 139722347643184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12], 139777662343520) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn, 139777662344720) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k, 139728976964640) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A, 139728976963872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'], 139722347644144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['second'].weight, 139722618619200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B, 139728976954656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B['second'], 139722347643808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.base_layer, 139777662344864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout, 139728976961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['second'], 139722347643376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q, 139728972691392) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A, 139728976524768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'], 139722347645296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['second'].weight, 139722618625760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B, 139728976517616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B['second'], 139722347645632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.base_layer, 139777662344960) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout, 139728976524384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['second'], 139722347645824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v, 139728976959264) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A, 139728976959168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'], 139722347495328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['second'].weight, 139722618630320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B, 139728976957776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B['second'], 139722347494272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.base_layer, 139777662345008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout, 139728976960512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['second'], 139722347643568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k, 139777662344912) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.weight, 139777664575232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q, 139777662344768) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.weight, 139781198835088) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.processor, 139777662344672) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm, 139777662344336) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm, 139777662344480) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu, 139777662344384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear, 139728971163312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A, 139728972861584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'], 139722347647072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['second'].weight, 139722625274064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B, 139728972869984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B['second'], 139722347645872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.base_layer, 139777662344432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout, 139728971169072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['second'], 139722347646688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp, 139777662344576) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp, 139728972862976) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A, 139728978267248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'], 139722347649136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['second'].weight, 139722618629120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B, 139728978270704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B['second'], 139722347644192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.base_layer, 139777662344528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout, 139728972863120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['second'], 139722347643664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out, 139728978280400) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A, 139728978277904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['second'], 139722347645920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['second'].weight, 139722618626640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B, 139728978268976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B['second'], 139722347646256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.base_layer, 139777662344624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout, 139728978274736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout['second'], 139722347646064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13], 139777662344288) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn, 139777662345488) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k, 139728976676128) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A, 139728976687312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'], 139722347479440) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['second'].weight, 139722618627680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B, 139728976682896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B['second'], 139722347480688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.base_layer, 139777662345632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout, 139728976675408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['second'], 139722347479536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q, 139728978261136) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A, 139728978256720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'], 139722347480016) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['second'].weight, 139722618627920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B, 139728978254032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B['second'], 139722347481024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.base_layer, 139777662345728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout, 139728978258016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['second'], 139722347483808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v, 139728977839920) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A, 139728977850768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'], 139722348667152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['second'].weight, 139722618622400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B, 139728977849424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B['second'], 139722348665088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.base_layer, 139777662345776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout, 139728977850432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['second'], 139722347479344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k, 139777662345680) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.weight, 139777664581312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q, 139777662345536) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.weight, 139777659801808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.processor, 139777662345440) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm, 139777662345104) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm, 139777662345248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu, 139777662345152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear, 139728976958160) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A, 139728976958640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'], 139722347494608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['second'].weight, 139722618618960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B, 139728976956144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B['second'], 139722347482800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.base_layer, 139777662345200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout, 139728976957248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['second'], 139722347482464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp, 139777662345344) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp, 139728976968096) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A, 139728976967808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'], 139722347490288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['second'].weight, 139722618626880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B, 139728976968528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B['second'], 139722347494368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.base_layer, 139777662345296) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout, 139728976953600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['second'], 139722347488368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out, 139728976969152) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A, 139728976965744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['second'], 139722347486256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['second'].weight, 139722618622880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B, 139728978254800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B['second'], 139722347480352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.base_layer, 139777662345392) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout, 139728976954416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout['second'], 139722347480496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14], 139777662345056) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn, 139777662346256) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k, 139728976866912) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A, 139728976857312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'], 139722349975472) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['second'].weight, 139722618627200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B, 139728976861920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B['second'], 139722349983824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.base_layer, 139777662346400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout, 139728976866528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['second'], 139722349985264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q, 139728977057712) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A, 139728977062128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'], 139722349984112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['second'].weight, 139722618626080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B, 139728977064912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B['second'], 139722349980800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.base_layer, 139777662346496) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout, 139728977052288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['second'], 139722349681952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v, 139728976856304) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A, 139728976856928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'], 139722349974368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['second'].weight, 139722618621840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B, 139728976855824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B['second'], 139722349973840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.base_layer, 139777662346544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout, 139728976858080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['second'], 139722349976288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k, 139777662346448) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.weight, 139777664751776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q, 139777662346304) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.weight, 139777664573952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.processor, 139777662346208) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm, 139777662345872) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm, 139777662346016) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu, 139777662345920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear, 139728978363056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A, 139728978365984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'], 139722352007712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['second'].weight, 139722618632080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B, 139728978376064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B['second'], 139722352013040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.base_layer, 139777662345968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout, 139728978373424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['second'], 139722348674784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp, 139777662346112) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp, 139728978369248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A, 139728978363008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'], 139722349554288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['second'].weight, 139722618626320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B, 139728978371024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B['second'], 139722349546992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.base_layer, 139777662346064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout, 139728978373568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['second'], 139722352005216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out, 139728978271616) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A, 139728977064672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['second'], 139722349689008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['second'].weight, 139722618621520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B, 139728977054832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B['second'], 139722349680368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.base_layer, 139777662346160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout, 139728977052864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout['second'], 139722349829552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15], 139777662345824) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn, 139777662347024) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k, 139728976897328) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A, 139728976891760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'], 139722350125184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['second'].weight, 139722625658896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B, 139728976895024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B['second'], 139722350121392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.base_layer, 139777662347168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout, 139728976897280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['second'], 139722350124704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q, 139728976717392) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A, 139728976716528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'], 139722350128736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['second'].weight, 139722625658736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B, 139728976717584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B['second'], 139722350127872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.base_layer, 139777662347264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout, 139728976719936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['second'], 139722350132672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v, 139728978363584) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A, 139728977147120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'], 139722350131280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['second'].weight, 139722625661056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B, 139728977145968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B['second'], 139722350124848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.base_layer, 139777662347312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout, 139728977147264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['second'], 139722350131520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k, 139777662347216) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.weight, 139777659855920) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q, 139777662347072) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.weight, 139777660118384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.processor, 139777662346976) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm, 139777662346640) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm, 139777662346784) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu, 139777662346688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear, 139728976862928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A, 139728976867056) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'], 139722350126864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['second'].weight, 139722623881344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B, 139728976862016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B['second'], 139722350129936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.base_layer, 139777662346736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout, 139728976866816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['second'], 139722349978352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp, 139777662346880) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp, 139728976866864) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A, 139728976855440) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'], 139722350127488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['second'].weight, 139722625658016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B, 139728976759488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B['second'], 139722350117552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.base_layer, 139777662346832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout, 139728976856256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['second'], 139722350117456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out, 139728976761504) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A, 139728976765152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['second'], 139722350127968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['second'].weight, 139722625652736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B, 139728976760208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B['second'], 139722350125760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.base_layer, 139777662346928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout, 139728976772832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout['second'], 139722350126048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16], 139777662346592) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn, 139777662347792) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k, 139728977139200) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A, 139728977138624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'], 139722349372816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['second'].weight, 139722625662576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B, 139728977146208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B['second'], 139722349373104) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.base_layer, 139777662347936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout, 139728977140352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['second'], 139722349372768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q, 139728977143232) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A, 139728977145056) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'], 139722349373392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['second'].weight, 139722625654016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B, 139728977142320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B['second'], 139722349373488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.base_layer, 139777662348032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout, 139728977140880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['second'], 139722349373248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v, 139728977090576) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A, 139728977091632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'], 139722349372336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['second'].weight, 139722623745328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B, 139728977097056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B['second'], 139722349372000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.base_layer, 139777662348080) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout, 139728977092496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['second'], 139722349372048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k, 139777662347984) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.weight, 139777664323712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q, 139777662347840) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.weight, 139777664746576) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.processor, 139777662347744) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm, 139777662347408) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm, 139777662347552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu, 139777662347456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear, 139728977148032) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A, 139728977147936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'], 139722349378528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['second'].weight, 139722625662336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B, 139728977140784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B['second'], 139722349379488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.base_layer, 139777662347504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout, 139728977139248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['second'], 139722350124032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp, 139777662347648) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp, 139728977146544) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A, 139728977144624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'], 139722349374928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['second'].weight, 139722625646816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B, 139728977146064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B['second'], 139722349374784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.base_layer, 139777662347600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout, 139728977148848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['second'], 139722349374688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out, 139728977141264) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A, 139728977139680) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['second'], 139722349373584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['second'].weight, 139722625659056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B, 139728977139488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B['second'], 139722349374160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.base_layer, 139777662347696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout, 139728977137904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout['second'], 139722349373296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17], 139777662347360) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn, 139777662348560) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k, 139728976902608) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A, 139728984827088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'], 139722349367824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['second'].weight, 139722618418368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B, 139728984821040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B['second'], 139722349368880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.base_layer, 139777662348704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout, 139728984821952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['second'], 139722349368976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q, 139728977241440) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A, 139728977241584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'], 139722349369456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['second'].weight, 139722618416688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B, 139728977247056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B['second'], 139722349369120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.base_layer, 139777662348800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout, 139728977239856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['second'], 139722349369168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v, 139728984951008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A, 139728984950480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'], 139722349367680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['second'].weight, 139722618419008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B, 139728984961616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B['second'], 139722349368208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.base_layer, 139777662348848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout, 139728984962528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['second'], 139722349367392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k, 139777662348752) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.weight, 139777660124544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q, 139777662348608) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.weight, 139777659810208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.processor, 139777662348512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm, 139777662348176) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm, 139777662348320) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu, 139777662348224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear, 139728977091920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A, 139728977089664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'], 139722349370704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['second'].weight, 139722623744608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B, 139728977089856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B['second'], 139722349371760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.base_layer, 139777662348272) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout, 139728977086112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['second'], 139722349371856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp, 139777662348416) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp, 139728977092400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A, 139728977089136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'], 139722349370560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['second'].weight, 139722621278608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B, 139728977093696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B['second'], 139722349371088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.base_layer, 139777662348368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout, 139728977096960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['second'], 139722349370368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out, 139728977238848) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A, 139728977241008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['second'], 139722349369840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['second'].weight, 139722618414848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B, 139728977238272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B['second'], 139722349370176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.base_layer, 139777662348464) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout, 139728977238944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout['second'], 139722349369648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18], 139777662348128) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn, 139777662349328) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k, 139728977232752) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A, 139728985170096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'], 139722349363888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['second'].weight, 139722617844608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B, 139728985172784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B['second'], 139722349364416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.base_layer, 139777662349472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout, 139728985057808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['second'], 139722349364224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q, 139728985001600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A, 139728985003952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'], 139722349364368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['second'].weight, 139722623920912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B, 139728985013456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B['second'], 139722349365088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.base_layer, 139777662349568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout, 139728985011248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['second'], 139722349364128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v, 139728985237456) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A, 139728985237648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'], 139722349363696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['second'].weight, 139722617831648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B, 139728985228096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B['second'], 139722349363408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.base_layer, 139777662349616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout, 139728985239040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['second'], 139722349363648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k, 139777662349520) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.weight, 139777660373728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q, 139777662349376) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.weight, 139777660372208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.processor, 139777662349280) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm, 139777662348944) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm, 139777662349088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu, 139777662348992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear, 139728984176528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A, 139728984167408) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'], 139722349366912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['second'].weight, 139722618419088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B, 139728984164576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B['second'], 139722349367152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.base_layer, 139777662349040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout, 139728984170336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['second'], 139722349367104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp, 139777662349184) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp, 139728985421472) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A, 139728985413264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'], 139722349365808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['second'].weight, 139722618412848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B, 139728985412448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B['second'], 139722349366384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.base_layer, 139777662349136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout, 139728985420512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['second'], 139722349366192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out, 139728985413552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A, 139728985412016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['second'], 139722349365232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['second'].weight, 139722618413888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B, 139728985421232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B['second'], 139722349365904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.base_layer, 139777662349232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout, 139728985409040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout['second'], 139722349365760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19], accessed_by=GetItemGuardAccessor(19) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19], 139777662348896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn, 139777662350096) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k, 139728983778512) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A, 139728983777168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'], 139722351352688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['second'].weight, 139722622860192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B, 139728983772560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B['second'], 139722351354224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.base_layer, 139777662350240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout, 139728983776688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['second'], 139722351352784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q, 139728985529056) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A, 139728985538224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'], 139722351352304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['second'].weight, 139722622425408) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B, 139728983785088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B['second'], 139722351349904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.base_layer, 139777662350336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout, 139728985534384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['second'], 139722351350912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v, 139728985587472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A, 139728985577920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'], 139722351359168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['second'].weight, 139722623269392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B, 139728985571776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B['second'], 139722351359936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.base_layer, 139777662350384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout, 139728985584640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['second'], 139722351360032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k, 139777662350288) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.weight, 139777664712304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q, 139777662350144) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.weight, 139777660377488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.processor, 139777662350048) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm, 139777662349712) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm, 139777662349856) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu, 139777662349760) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear, 139728985268016) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A, 139728985267248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'], 139722351361904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['second'].weight, 139722617836768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B, 139728985271568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B['second'], 139722351360512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.base_layer, 139777662349808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout, 139728985264512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['second'], 139722351361856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp, 139777662349952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp, 139728985378528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A, 139728985482976) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'], 139722351360320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['second'].weight, 139722622754768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B, 139728985483936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B['second'], 139722351361184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.base_layer, 139777662349904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout, 139728985390288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['second'], 139722351361280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out, 139728985473616) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A, 139728985479184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['second'], 139722351347792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['second'].weight, 139722622761088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B, 139728985531264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B['second'], 139722351352016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.base_layer, 139777662350000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout, 139728985482832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout['second'], 139722351351392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20], accessed_by=GetItemGuardAccessor(20) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20], 139777662349664) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn, 139777662350864) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k, 139728985675840) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A, 139728985674784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'], 139722351355376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['second'].weight, 139722618922032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B, 139728985681696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B['second'], 139722351356096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.base_layer, 139777662351008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout, 139728985676128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['second'], 139722351355280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q, 139728985775152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A, 139728985772128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'], 139722351358400) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['second'].weight, 139722618927872) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B, 139728985783840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B['second'], 139722351353024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.base_layer, 139777662351104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout, 139728985776448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['second'], 139722351358496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v, 139728985014224) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A, 139729027839504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'], 139722351355136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['second'].weight, 139722622238864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B, 139729027836960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B['second'], 139722351354512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.base_layer, 139777662351152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout, 139729027839168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['second'], 139722351354656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k, 139777662351056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.weight, 139777664182496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q, 139777662350912) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.weight, 139777664729392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.processor, 139777662350816) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm, 139777662350480) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm, 139777662350624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu, 139777662350528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear, 139728984115040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A, 139728984117104) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'], 139722351357536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['second'].weight, 139722621877712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B, 139728984115184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B['second'], 139722351358064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.base_layer, 139777662350576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout, 139728984118496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['second'], 139722351358160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp, 139777662350720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp, 139728984121712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A, 139728984121424) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'], 139722351356768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['second'].weight, 139722621865552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B, 139728984127856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B['second'], 139722351357344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.base_layer, 139777662350672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout, 139728984123440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['second'], 139722351357296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out, 139728985725904) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A, 139728985726672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['second'], 139722351356384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['second'].weight, 139722621022544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B, 139728985730080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B['second'], 139722351356048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.base_layer, 139777662350768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout, 139728985725760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout['second'], 139722351356192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21], accessed_by=GetItemGuardAccessor(21) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21], 139777662350432) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn, 139777662351632) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k, 139729028935408) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A, 139729071121568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'], 139722351194032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['second'].weight, 139722621123088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B, 139729071115664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B['second'], 139722351194368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.base_layer, 139777662351776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout, 139729071261584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['second'], 139722351197872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q, 139729027941952) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A, 139729070046176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'], 139722351187648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['second'].weight, 139722621121568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B, 139729071757088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B['second'], 139722351186832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.base_layer, 139777662351872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout, 139729070400864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['second'], 139722351185824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v, 139729071458816) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A, 139729073486720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'], 139722351195616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['second'].weight, 139722621321136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B, 139729073485520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B['second'], 139722351193552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.base_layer, 139777662351920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout, 139729072770544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['second'], 139722351193840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k, 139777662351824) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.weight, 139777660377808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q, 139777662351680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.weight, 139777664725328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.processor, 139777662351584) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm, 139777662351248) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm, 139777662351392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu, 139777662351296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear, 139729031521440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A, 139729030869728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'], 139722351353216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['second'].weight, 139722620421920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B, 139729073071840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B['second'], 139722351353264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.base_layer, 139777662351344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout, 139729030517024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['second'], 139722351353504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp, 139777662351488) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp, 139729029447152) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A, 139729029435152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'], 139722351352496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['second'].weight, 139722619279920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B, 139729029300480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B['second'], 139722351351248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.base_layer, 139777662351440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout, 139729029446864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['second'], 139722352155552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out, 139729029021120) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A, 139729029123808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['second'], 139722351189376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['second'].weight, 139722623792000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B, 139729029121936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B['second'], 139722351192880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.base_layer, 139777662351536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout, 139729029132256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout['second'], 139722351192928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22], accessed_by=GetItemGuardAccessor(22) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22], 139777662351200) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn, 139777662352400) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k, 139729075138144) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A, 139729075128688) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'], 139722350549200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['second'].weight, 139722623634304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B, 139729076787024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B['second'], 139722350553184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.base_layer, 139777662352544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout, 139729075138480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['second'], 139722352486064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q, 139729073604208) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A, 139729075662192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'], 139722351084208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['second'].weight, 139722620243136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B, 139729075661328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B['second'], 139722351092416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.base_layer, 139777662352640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout, 139729075660608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['second'], 139722351093664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v, 139729078054256) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A, 139729078057568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'], 139722350555008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['second'].weight, 139722621624112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B, 139729078050320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B['second'], 139722350553136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.base_layer, 139777662352688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout, 139729078055120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['second'], 139722350547808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k, 139777662352592) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.weight, 139777664757856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q, 139777662352448) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.weight, 139777660378528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.processor, 139777662352352) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm, 139777662352016) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm, 139777662352160) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu, 139777662352064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear, 139729072306688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A, 139729072306448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'], 139722351189184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['second'].weight, 139722622145360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B, 139729072306208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B['second'], 139722351189712) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.base_layer, 139777662352112) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout, 139729072294208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['second'], 139722351183280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp, 139777662352256) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp, 139729072038160) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A, 139729071776400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'], 139722351187696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['second'].weight, 139722621555296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B, 139729071770016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B['second'], 139722351187024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.base_layer, 139777662352208) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout, 139729071783648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['second'], 139722351186640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out, 139729070823680) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A, 139729073304432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['second'], 139722351094624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['second'].weight, 139722619058144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B, 139729073304528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B['second'], 139722351096256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.base_layer, 139777662352304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout, 139729073299296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout['second'], 139722351182704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23], accessed_by=GetItemGuardAccessor(23) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23], 139777662351968) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn, 139777662353168) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k, 139729076457856) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A, 139729076456224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'], 139722351585856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['second'].weight, 139722621731760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B, 139729076457232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B['second'], 139722351582928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.base_layer, 139777661026368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout, 139729076466736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['second'], 139722351589504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q, 139729074952976) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A, 139729074950432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'], 139722351579280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['second'].weight, 139722620924160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B, 139729074946496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B['second'], 139722351588688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.base_layer, 139777662353312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout, 139729074961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['second'], 139722351468768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v, 139729076356960) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A, 139729077922128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'], 139722351725536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['second'].weight, 139722622350288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B, 139729077927504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B['second'], 139722351731680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.base_layer, 139777661026416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout, 139729077912144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['second'], 139722351591040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k, 139777662353360) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.weight, 139777660445824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q, 139777662353216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.weight, 139777664332992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.processor, 139777662353120) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm, 139777662352784) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm, 139777662352928) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu, 139777662352832) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear, 139728986592768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A, 139728986129744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'], 139722353763776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['second'].weight, 139722622383776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B, 139728986142176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B['second'], 139722353766416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.base_layer, 139777662352880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout, 139728986590560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['second'], 139722350555584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp, 139777662353024) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp, 139728986134400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A, 139728986136272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'], 139722353669264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['second'].weight, 139722621219152) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B, 139728986134736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B['second'], 139722353667344) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.base_layer, 139777662352976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout, 139728986133104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['second'], 139722353668304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out, 139728986503344) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A, 139728986498544) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['second'], 139722353368256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['second'].weight, 139722622039072) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B, 139729076928080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B['second'], 139722353376560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.base_layer, 139777662353072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout, 139728986495088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout['second'], 139722353894272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24], accessed_by=GetItemGuardAccessor(24) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24], 139777662352736) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn, 139777661026896) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k, 139729080562160) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A, 139729080563120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'], 139722351827872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['second'].weight, 139722617271008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B, 139729080563600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B['second'], 139722351833968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.base_layer, 139777661027040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout, 139729080562592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['second'], 139722351734608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q, 139729079640528) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A, 139729079641824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'], 139722351733024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['second'].weight, 139722617272608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B, 139729079632080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B['second'], 139722351737056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.base_layer, 139777661027136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout, 139729079637408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['second'], 139722351736240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v, 139729080553184) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A, 139729080557936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'], 139722351831280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['second'].weight, 139722617265488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B, 139729080552992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B['second'], 139722351830848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.base_layer, 139777661027184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout, 139729080562496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['second'], 139722351831520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k, 139777661027088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.weight, 139777660322976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q, 139777661026944) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.weight, 139777664578432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.processor, 139777661026848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm, 139777661026512) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm, 139777661026656) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu, 139777661026560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear, 139729077811472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A, 139729077580800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'], 139722351728080) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['second'].weight, 139722622339088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B, 139729077576864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B['second'], 139722351727648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.base_layer, 139777661026608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout, 139729077812960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['second'], 139722351727888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp, 139777661026752) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp, 139729079646912) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A, 139729079643840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'], 139722351736672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['second'].weight, 139722623953680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B, 139729079638608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B['second'], 139722351736624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.base_layer, 139777661026704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout, 139729079647344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['second'], 139722351737536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out, 139729079643936) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A, 139729079634960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['second'], 139722351735856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['second'].weight, 139722624043424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B, 139729079635728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B['second'], 139722351738304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.base_layer, 139777661026800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout, 139729079646384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout['second'], 139722351734368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25], accessed_by=GetItemGuardAccessor(25) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25], 139777661026464) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn, 139777661027664) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k, 139729081832768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A, 139729081831280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'], 139722353083680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['second'].weight, 139722617271248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B, 139729081840016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B['second'], 139722353083200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.base_layer, 139777661027808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout, 139729081828112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['second'], 139722353180880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q, 139729081832816) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A, 139729081828352) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'], 139722353181216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['second'].weight, 139722617273888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B, 139729081828640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B['second'], 139722353182608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.base_layer, 139777661027904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout, 139729081835792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['second'], 139722353192880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v, 139729082854144) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A, 139729082857168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'], 139722353097408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['second'].weight, 139722617263888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B, 139729082857936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B['second'], 139722353098320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.base_layer, 139777661027952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout, 139729082855536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['second'], 139722353098176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k, 139777661027856) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.weight, 139777664182896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q, 139777661027712) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.weight, 139777664566032) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.processor, 139777661027616) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm, 139777661027280) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm, 139777661027424) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu, 139777661027328) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear, 139729080550784) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A, 139729080552368) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'], 139722351829984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['second'].weight, 139722617264048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B, 139729080550736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B['second'], 139722351830896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.base_layer, 139777661027376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout, 139729080549728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['second'], 139722351831136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp, 139777661027520) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp, 139729080556496) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A, 139729081837952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'], 139722351829792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['second'].weight, 139722617258048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B, 139729081834208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B['second'], 139722351822112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.base_layer, 139777661027472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout, 139729080553328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['second'], 139722351829696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out, 139729081843424) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A, 139729081839584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['second'], 139722351830704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['second'].weight, 139722617260928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B, 139729081837040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B['second'], 139722351821632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.base_layer, 139777661027568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout, 139729081840832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout['second'], 139722351824272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26], accessed_by=GetItemGuardAccessor(26) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26], 139777661027232) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn, 139777661028432) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k, 139729085484240) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A, 139729085485200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'], 139722353094480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['second'].weight, 139722617262128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B, 139729085485680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B['second'], 139722353095152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.base_layer, 139777661028576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout, 139729085484672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['second'], 139722353095248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q, 139729085484576) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A, 139729085491008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'], 139722353095968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['second'].weight, 139722617271328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B, 139729085491488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B['second'], 139722353095680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.base_layer, 139777661028672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout, 139729085493600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['second'], 139722353095920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v, 139729085490720) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A, 139729085481072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'], 139722353093760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['second'].weight, 139722617269008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B, 139729085490288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B['second'], 139722353094144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.base_layer, 139777661028720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout, 139729085490624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['second'], 139722353093520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k, 139777661028624) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.weight, 139777661092384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q, 139777661028480) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.weight, 139777661093744) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.processor, 139777661028384) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm, 139777661028048) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm, 139777661028192) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu, 139777661028096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear, 139729082856928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A, 139729082852752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'], 139722353097360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['second'].weight, 139722617260688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B, 139729082850496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B['second'], 139722353097888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.base_layer, 139777661028144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout, 139729082857264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['second'], 139722353097840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp, 139777661028288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp, 139729082848528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A, 139729082843872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'], 139722353096880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['second'].weight, 139722617258608) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B, 139729082844112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B['second'], 139722353097264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.base_layer, 139777661028240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout, 139729082850304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['second'], 139722353096400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out, 139729082848720) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A, 139729082857840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['second'], 139722353096160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['second'].weight, 139722617273328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B, 139729085496000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B['second'], 139722353096688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.base_layer, 139777661028336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout, 139729082855920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout['second'], 139722353096496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27], accessed_by=GetItemGuardAccessor(27) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27], 139777661028000) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn, 139777661029200) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k, 139729087700448) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A, 139729087699008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'], 139722353090304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['second'].weight, 139722626030048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B, 139729087699296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B['second'], 139722353090928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.base_layer, 139777661029344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout, 139729087704720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['second'], 139722353090544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q, 139729073171488) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A, 139729087694208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'], 139722353091216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['second'].weight, 139722617258528) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B, 139729087698864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B['second'], 139722353091312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.base_layer, 139777661029440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout, 139729087703568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['second'], 139722353090832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v, 139729087693392) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A, 139729087693536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'], 139722353089872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['second'].weight, 139722626025248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B, 139729087693296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B['second'], 139722353089296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.base_layer, 139777661029488) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout, 139729087693008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['second'], 139722353089824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k, 139777661029392) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.weight, 139777661093504) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q, 139777661029248) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.weight, 139777661094864) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.processor, 139777661029152) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm, 139777661028816) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm, 139777661028960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu, 139777661028864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear, 139729086476304) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A, 139729086478752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'], 139722353093280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['second'].weight, 139722617267648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B, 139729086478608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B['second'], 139722353093808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.base_layer, 139777661028912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout, 139729086477696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['second'], 139722353093616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp, 139777661029056) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp, 139729086469584) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A, 139729086471840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'], 139722353093088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['second'].weight, 139722617263088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B, 139729086466560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B['second'], 139722353092800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.base_layer, 139777661029008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout, 139729086472224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['second'], 139722353093040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out, 139729086466416) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A, 139729086468816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['second'], 139722353091504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['second'].weight, 139722617258448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B, 139729086469248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B['second'], 139722353092272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.base_layer, 139777661029104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout, 139729086464112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout['second'], 139722353092368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28], accessed_by=GetItemGuardAccessor(28) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28], 139777661028768) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn, 139777661030016) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k, 139729091443536) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A, 139729092148384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'], 139722353085696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['second'].weight, 139722626035648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B, 139729092139648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B['second'], 139722353086176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.base_layer, 139777661030160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout, 139729092142528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['second'], 139722353086128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q, 139729091429616) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A, 139729091429280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'], 139722353087088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['second'].weight, 139722626024128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B, 139729091431680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B['second'], 139722353086704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.base_layer, 139777661030256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout, 139729091429712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['second'], 139722353087040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v, 139729092148624) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A, 139729092140992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'], 139722353085456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['second'].weight, 139722626034688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B, 139729092137152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B['second'], 139722353085552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.base_layer, 139777661030304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout, 139729092143968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['second'], 139722353084928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k, 139777661030208) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.weight, 139777661094624) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q, 139777661030064) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.weight, 139777661096064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.processor, 139777661029968) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm, 139777661029584) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm, 139777661029728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu, 139777661029632) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear, 139729087705872) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A, 139729087707792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'], 139722353088240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['second'].weight, 139722626034128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B, 139729087708272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B['second'], 139722353089152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.base_layer, 139777661029680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout, 139729087695792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['second'], 139722353089248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp, 139777661029872) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp, 139729091442720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A, 139729091439984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'], 139722353088336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['second'].weight, 139722626024928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B, 139729091439888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B['second'], 139722353088384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.base_layer, 139777661029776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout, 139729091442768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['second'], 139722353087952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out, 139729091434944) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A, 139729091435040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['second'], 139722353087760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['second'].weight, 139722626033968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B, 139729091434368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B['second'], 139722353087280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.base_layer, 139777661029920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout, 139729091437056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout['second'], 139722353087712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29], accessed_by=GetItemGuardAccessor(29) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29], 139777661029536) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn, 139777661030784) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k, 139729093490768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A, 139729093490144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'], 139722352950960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['second'].weight, 139722626028448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B, 139729093486544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B['second'], 139722352946784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.base_layer, 139777661030928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout, 139729093483760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['second'], 139722352947168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q, 139729093484624) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A, 139729093491872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'], 139722352946496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['second'].weight, 139722626030208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B, 139729093484768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B['second'], 139722352940832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.base_layer, 139777661031024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout, 139729093482800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['second'], 139722352946832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v, 139729093483472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A, 139729093484576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'], 139722352950576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['second'].weight, 139722626036768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B, 139729093481744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B['second'], 139722352950240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.base_layer, 139777661031072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout, 139729093481984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['second'], 139722352950384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k, 139777661030976) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.weight, 139777661095824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q, 139777661030832) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.weight, 139777661097184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.processor, 139777661030736) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm, 139777661030400) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm, 139777661030544) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu, 139777661030448) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear, 139729092135568) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A, 139729092139744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'], 139722353084832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['second'].weight, 139722626030128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B, 139729092141376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B['second'], 139722353085264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.base_layer, 139777661030496) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout, 139729092133264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['second'], 139722353084784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp, 139777661030640) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp, 139729092147040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A, 139729092147280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'], 139722353082912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['second'].weight, 139722626026768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B, 139729092137824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B['second'], 139722353083488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.base_layer, 139777661030592) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout, 139729092141808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['second'], 139722353082960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out, 139729092133552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A, 139729092134800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['second'], 139722352943712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['second'].weight, 139722626027008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B, 139729092132928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B['second'], 139722352950816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.base_layer, 139777661030688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout, 139729092133648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout['second'], 139722353083104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30], accessed_by=GetItemGuardAccessor(30) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30], 139777661030352) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn, 139777661031552) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k, 139729091698624) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A, 139729091700928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'], 139722352941360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['second'].weight, 139722626030688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B, 139729091692672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B['second'], 139722352942272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.base_layer, 139777661031696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout, 139729091697376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['second'], 139722352938144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q, 139729091695024) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A, 139729091697232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'], 139722352947120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['second'].weight, 139722626028928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B, 139729091696752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B['second'], 139722352947696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.base_layer, 139777661031792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout, 139729091695120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['second'], 139722352947312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v, 139729093047056) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A, 139729093043168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'], 139722352935600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['second'].weight, 139722626032368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B, 139729093043072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B['second'], 139722352935312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.base_layer, 139777661031840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout, 139729093047680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['second'], 139722352945152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k, 139777661031744) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.weight, 139777661096944) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q, 139777661031600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.weight, 139777661098304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.processor, 139777661031504) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm, 139777661031168) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm, 139777661031312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu, 139777661031216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear, 139729093480016) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A, 139729093480976) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'], 139722352948464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['second'].weight, 139722626028208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B, 139729093481024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B['second'], 139722352947216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.base_layer, 139777661031264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout, 139729093480064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['second'], 139722352947264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp, 139777661031408) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp, 139729091705296) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A, 139729091703232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'], 139722352949280) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['second'].weight, 139722626038368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B, 139729091699968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B['second'], 139722352949472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.base_layer, 139777661031360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout, 139729091706736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['second'], 139722352949520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out, 139729091701456) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A, 139729091691472) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['second'], 139722352948608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['second'].weight, 139722626038048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B, 139729091690848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B['second'], 139722352948848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.base_layer, 139777661031456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout, 139729091700592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout['second'], 139722352948560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31], accessed_by=GetItemGuardAccessor(31) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31], 139777661031120) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn, 139777661032320) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k, 139729093175344) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A, 139729093165984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'], 139722352829552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['second'].weight, 139722625218208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B, 139729093170736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B['second'], 139722352833200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.base_layer, 139777661032464) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout, 139729093166224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['second'], 139722352830176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q, 139729093177696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A, 139729093178656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'], 139722352832720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['second'].weight, 139722625213088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B, 139729093179136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B['second'], 139722352836368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.base_layer, 139777661032560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout, 139729093178128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['second'], 139722352832624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v, 139729093173328) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A, 139729093165696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'], 139722352830944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['second'].weight, 139722625218768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B, 139729093170448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B['second'], 139722352829984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.base_layer, 139777661032608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout, 139729093172944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['second'], 139722352596960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k, 139777661032512) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.weight, 139777661098064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q, 139777661032368) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.weight, 139777661099424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.processor, 139777661032272) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm, 139777661031936) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm, 139777661032080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu, 139777661031984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear, 139729093035440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A, 139729093035200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'], 139722352833344) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['second'].weight, 139722626036128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B, 139729093034480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B['second'], 139722352835072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.base_layer, 139777661032032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout, 139729093043504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['second'], 139722352834976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp, 139777661032176) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp, 139729093040624) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A, 139729093037072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'], 139722352834496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['second'].weight, 139722625219248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B, 139729093040816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B['second'], 139722352822688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.base_layer, 139777661032128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout, 139729093039328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['second'], 139722352832864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out, 139729093035920) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A, 139729093172080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['second'], 139722352833056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['second'].weight, 139722625215888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B, 139729093168672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B['second'], 139722352830368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.base_layer, 139777661032224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout, 139729093046768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout['second'], 139722352834928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32], accessed_by=GetItemGuardAccessor(32) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32], 139777661031888) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn, 139777661033088) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k, 139729094389680) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A, 139729094390640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'], 139722355000368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['second'].weight, 139722625216368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B, 139729094393568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B['second'], 139722355003920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.base_layer, 139777661033232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout, 139729094391600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['second'], 139722357676624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q, 139729093649472) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A, 139729093650672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'], 139722354411984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['second'].weight, 139722625215168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B, 139729093641648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B['second'], 139722354412080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.base_layer, 139777661033328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout, 139729093649904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['second'], 139722354413280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v, 139729094381376) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A, 139729094387328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'], 139722355148592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['second'].weight, 139722625835200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B, 139729094383008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B['second'], 139722355155888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.base_layer, 139777661033376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout, 139729094381808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['second'], 139722355004784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k, 139777661033280) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.weight, 139777661099184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q, 139777661033136) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.weight, 139777661100544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.processor, 139777661033040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm, 139777661032704) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm, 139777661032848) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu, 139777661032752) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear, 139729093655856) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A, 139729093652448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'], 139722352696704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['second'].weight, 139722625212048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B, 139729093653264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B['second'], 139722352704048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.base_layer, 139777661032800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout, 139729093655808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['second'], 139722352704144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp, 139777661032944) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp, 139729093653168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A, 139729093647312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'], 139722354268080) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['second'].weight, 139722625205568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B, 139729093654032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B['second'], 139722354273648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.base_layer, 139777661032896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout, 139729093653024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['second'], 139722352690512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out, 139729093644240) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A, 139729093644912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['second'], 139722354413808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['second'].weight, 139722625204848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B, 139729093647552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B['second'], 139722354414480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.base_layer, 139777661032992) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout, 139729093644624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout['second'], 139722354267264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33], accessed_by=GetItemGuardAccessor(33) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33], 139777661032656) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn, 139777661033856) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k, 139729092741776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A, 139729092749456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'], 139722355425728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['second'].weight, 139722625833120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B, 139729092744224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B['second'], 139722355426208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.base_layer, 139777661034000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout, 139729092742736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['second'], 139722355440896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q, 139729092747008) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A, 139729092747392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'], 139722355439360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['second'].weight, 139722625841040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B, 139729092745280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B['second'], 139722355438496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.base_layer, 139777661034096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout, 139729092747536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['second'], 139722355440800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v, 139729092749264) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A, 139729094561552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'], 139722355441568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['second'].weight, 139722625829280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B, 139729094560544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B['second'], 139722355437536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.base_layer, 139777661034144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout, 139729094571344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['second'], 139722355432640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k, 139777661034048) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.weight, 139777661100304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q, 139777661033904) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.weight, 139777661101664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.processor, 139777661033808) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm, 139777661033472) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm, 139777661033616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu, 139777661033520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear, 139729094379024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A, 139729094381232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'], 139722355284800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['second'].weight, 139722625840880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B, 139729094380560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B['second'], 139722355286096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.base_layer, 139777661033568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout, 139729094379120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['second'], 139722355151136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp, 139777661033712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp, 139729094393808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A, 139729092755120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'], 139722355428416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['second'].weight, 139722625838400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B, 139729092755168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B['second'], 139722355438208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.base_layer, 139777661033664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout, 139729092750752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['second'], 139722355292816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out, 139729092754064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A, 139729092742880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['second'], 139722355435664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['second'].weight, 139722625838800) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B, 139729092749792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B['second'], 139722355435616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.base_layer, 139777661033760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout, 139729092753440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout['second'], 139722355436576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34], accessed_by=GetItemGuardAccessor(34) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34], 139777661033424) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn, 139777661034624) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k, 139729095747776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A, 139729095745136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'], 139722355570240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['second'].weight, 139722625826960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B, 139729095742592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B['second'], 139722355571008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.base_layer, 139777661034768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout, 139729095747488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['second'], 139722355557712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q, 139729095749120) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A, 139729095753248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'], 139722355570912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['second'].weight, 139722625833440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B, 139729095739808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B['second'], 139722355570624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.base_layer, 139777661034864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout, 139729095751088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['second'], 139722355562656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v, 139729095737744) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A, 139729095739088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'], 139722355569808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['second'].weight, 139722625837600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B, 139729095738752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B['second'], 139722355563952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.base_layer, 139777661034912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout, 139729095743888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['second'], 139722355569904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k, 139777661034816) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.weight, 139777661101424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q, 139777661034672) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.weight, 139777661102784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.processor, 139777661034576) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm, 139777661034240) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm, 139777661034384) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu, 139777661034288) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear, 139729094571392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A, 139729094570528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'], 139722355564048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['second'].weight, 139722625831840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B, 139729094573456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B['second'], 139722355567792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.base_layer, 139777661034336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout, 139729094569712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['second'], 139722355432736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp, 139777661034480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp, 139729094566400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A, 139729094566304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'], 139722355572640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['second'].weight, 139722625827360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B, 139729094565056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B['second'], 139722355569040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.base_layer, 139777661034432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout, 139729094564720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['second'], 139722355567648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out, 139729094562416) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A, 139729094562704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['second'], 139722355572112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['second'].weight, 139722625841120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B, 139729094561936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B['second'], 139722355568800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.base_layer, 139777661034528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout, 139729094560880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout['second'], 139722355568752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35], accessed_by=GetItemGuardAccessor(35) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35], 139777661034192) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn, 139777661035392) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k, 139729093790256) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A, 139729095007568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'], 139722354821968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['second'].weight, 139722625828880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B, 139729095010448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B['second'], 139722354822832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.base_layer, 139777661035536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout, 139729093789728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['second'], 139722354822400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q, 139729093792176) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A, 139729093802736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'], 139722354823456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['second'].weight, 139722625835360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B, 139729093788768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B['second'], 139722354823024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.base_layer, 139777661035632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout, 139729093793280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['second'], 139722354823312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v, 139729095010880) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A, 139729095011984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'], 139722354822064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['second'].weight, 139722625832720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B, 139729095012224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B['second'], 139722354822256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.base_layer, 139777661035680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout, 139729095007280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['second'], 139722354821680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k, 139777661035584) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.weight, 139777661102544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q, 139777661035440) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.weight, 139777661103904) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.processor, 139777661035344) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm, 139777661035008) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm, 139777661035152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu, 139777661035056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear, 139729095750512) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A, 139729095745376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'], 139722355564816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['second'].weight, 139722625828800) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B, 139729095747056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B['second'], 139722355571200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.base_layer, 139777661035104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout, 139729095742208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['second'], 139722355565440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp, 139777661035248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp, 139729093800768) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A, 139729093803024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'], 139722354824608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['second'].weight, 139722625834560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B, 139729093797216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B['second'], 139722354832528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.base_layer, 139777661035200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout, 139729093798416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['second'], 139722355557088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out, 139729093803072) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A, 139729093800720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['second'], 139722354823792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['second'].weight, 139722625831680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B, 139729093795872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B['second'], 139722354824368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.base_layer, 139777661035296) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout, 139729093801776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout['second'], 139722354824080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36], accessed_by=GetItemGuardAccessor(36) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36], 139777661034960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn, 139777661036160) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k, 139729098327424) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A, 139729098327088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'], 139722354736960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['second'].weight, 139722618604656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B, 139729098329632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B['second'], 139722354737008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.base_layer, 139777661036304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout, 139729098326176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['second'], 139722354736576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q, 139729098331264) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A, 139729098332080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'], 139722354727024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['second'].weight, 139722618605616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B, 139729098330016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B['second'], 139722354726544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.base_layer, 139777661036400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout, 139729098329872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['second'], 139722354819616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v, 139729098337696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A, 139729098789920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'], 139722354736384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['second'].weight, 139722618613296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B, 139729098799952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B['second'], 139722354735904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.base_layer, 139777661036448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout, 139729098335632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['second'], 139722354736336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k, 139777661036352) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.weight, 139777661103664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q, 139777661036208) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.weight, 139777661105024) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.processor, 139777661036112) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm, 139777661035776) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm, 139777661035920) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu, 139777661035824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear, 139729095003584) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A, 139729095006128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'], 139722354821152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['second'].weight, 139722625838160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B, 139729095004544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B['second'], 139722354821776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.base_layer, 139777661035872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout, 139729095004880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['second'], 139722354821392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp, 139777661036016) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp, 139729095000368) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A, 139729095001808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'], 139722354820720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['second'].weight, 139722625829440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B, 139729095003536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B['second'], 139722354820144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.base_layer, 139777661035968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout, 139729095002720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['second'], 139722354820672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out, 139729098339520) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A, 139729098338848) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['second'], 139722354819376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['second'].weight, 139722618610896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B, 139729098335536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B['second'], 139722354820000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.base_layer, 139777661036064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout, 139729098337840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout['second'], 139722354820096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37], accessed_by=GetItemGuardAccessor(37) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37], 139777661035728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn, 139777661036928) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k, 139729099849104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A, 139729099847808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'], 139722354732592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['second'].weight, 139722618612656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B, 139729099848576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B['second'], 139722354732208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.base_layer, 139777661037072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout, 139729099846224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['second'], 139722354732544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_k.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_k.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_k.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q, 139729098795152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A, 139729099837968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'], 139722354733168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['second'].weight, 139722618613536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B, 139729099842528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B['second'], 139722354733792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.base_layer, 139777661037168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout, 139729099843824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['second'], 139722354733360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_q.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_q.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_q.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v, 139729099838208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A, 139729099841472) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'], 139722354731296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['second'].weight, 139722618607056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B, 139729099840608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B['second'], 139722354731920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.base_layer, 139777661037216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout, 139729099835856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['second'], 139722354731872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_v.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_v.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_v.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k, 139777661037120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.weight, 139777661104784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q, 139777661036976) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.weight, 139777661106144) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.processor, 139777661036880) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm, 139777661036544) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm, 139777661036688) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu, 139777661036592) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear, 139729098795440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A, 139729098793952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'], 139722354735712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['second'].weight, 139722618616816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B, 139729098796208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B['second'], 139722354735328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.base_layer, 139777661036640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout, 139729098798560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['second'], 139722354735664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].norm.linear.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].norm.linear.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].norm.linear.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp, 139777661036784) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp, 139729098789440) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A, 139729098787040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'], 139722354734176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['second'].weight, 139722618617056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B, 139729098788288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B['second'], 139722354734464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.base_layer, 139777661036736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout, 139729098794528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['second'], 139722354734608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_mlp.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].proj_mlp.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_mlp.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out, 139729098790736) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A, 139729098796544) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['second'], 139722354733888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['second'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['second']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['second'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['second'].weight, 139722618615936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B, 139729098797312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B['second'], 139722354733984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B['second'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.base_layer, 139777661036832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout, 139729098795584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['second'], accessed_by=GetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout['second'], 139722354733840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['second'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['second'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout['second'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_out.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].proj_out.scaling['second'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_out.use_dora) == 2 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.use_dora['second'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['img_ids'], accessed_by=DictGetItemGuardAccessor(img_ids) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['img_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[4096, 3], stride=[3, 1]) # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['img_ids'], '_dynamo_dynamic_indices') == False # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['txt_ids'], accessed_by=DictGetItemGuardAccessor(txt_ids) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['txt_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[512, 3], stride=[3, 1]) # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['txt_ids'], '_dynamo_dynamic_indices') == False # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['guidance'], accessed_by=DictGetItemGuardAccessor(guidance) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['guidance'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.float32, device=0, requires_grad=False, size=[1], stride=[1]) # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['guidance'], '_dynamo_dynamic_indices') == False # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['timestep'], accessed_by=DictGetItemGuardAccessor(timestep) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['timestep'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1], stride=[1]) # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['timestep'], '_dynamo_dynamic_indices') == False # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['return_dict'], accessed_by=DictGetItemGuardAccessor(return_dict) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- ID_MATCH: ___check_obj_id(L['return_dict'], 139786895605728) # if not return_dict: # diffusers/src/diffusers/models/transformers/transformer_flux.py:555 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['hidden_states'], accessed_by=DictGetItemGuardAccessor(hidden_states) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 4096, 64], stride=[262144, 64, 1]) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['pooled_projections'], accessed_by=DictGetItemGuardAccessor(pooled_projections) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['pooled_projections'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 768], stride=[768, 1]) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['pooled_projections'], '_dynamo_dynamic_indices') == False # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['encoder_hidden_states'], accessed_by=DictGetItemGuardAccessor(encoder_hidden_states) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- TENSOR_MATCH: check_tensor(L['encoder_hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 512, 4096], stride=[2097152, 4096, 1]) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_HASATTR: hasattr(L['encoder_hidden_states'], '_dynamo_dynamic_indices') == False # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['joint_attention_kwargs'], accessed_by=DictGetItemGuardAccessor(joint_attention_kwargs) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- ID_MATCH: ___check_obj_id(L['joint_attention_kwargs'], 139786895690912) # if joint_attention_kwargs is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['controlnet_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_block_samples) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_block_samples'], 139786895690912) # if controlnet_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:502 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=L['controlnet_single_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_single_block_samples) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_single_block_samples'], 139786895690912) # if controlnet_single_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:538 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | +- GuardManager: source=G, accessed_by=GlobalsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['torch'], accessed_by=DictGetItemGuardAccessor(torch) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['torch'], 139786870267216) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['torch'].cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['torch'].cat, 139786867250080) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['torch'].float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- EQUALS_MATCH: G['torch'].float16 == torch.float16 # if encoder_hidden_states.dtype == torch.float16: # diffusers/src/diffusers/models/transformers/transformer_flux.py:200 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['USE_PEFT_BACKEND'], accessed_by=DictGetItemGuardAccessor(USE_PEFT_BACKEND) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['USE_PEFT_BACKEND'], 139786895605696) # if USE_PEFT_BACKEND: # diffusers/src/diffusers/models/transformers/transformer_flux.py:434 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['scale_lora_layers'], accessed_by=DictGetItemGuardAccessor(scale_lora_layers) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['scale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['scale_lora_layers'].__code__, 139780914343312) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['unscale_lora_layers'], accessed_by=DictGetItemGuardAccessor(unscale_lora_layers) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['unscale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['unscale_lora_layers'].__code__, 139780914343488) # unscale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:553 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__builtins_dict___4'], accessed_by=DictGetItemGuardAccessor(__builtins_dict___4) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['int'], accessed_by=DictGetItemGuardAccessor(int) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['int'], 139786895673536) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['len'], accessed_by=DictGetItemGuardAccessor(len) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['len'], 139786873098768) # assert len(timesteps.shape) == 1, "Timesteps should be a 1d-array" # diffusers/src/diffusers/models/embeddings.py:54 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['set'], accessed_by=DictGetItemGuardAccessor(set) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['set'], 139786895697184) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['str'], accessed_by=DictGetItemGuardAccessor(str) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['str'], 139786895715456) # if isinstance(self.active_adapter, str): # peft/tuners/tuners_utils.py:530 in active_adapters V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['bool'], accessed_by=DictGetItemGuardAccessor(bool) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['bool'], 139786895605760) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['range'], accessed_by=DictGetItemGuardAccessor(range) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['range'], 139786895693664) # for i in range(n_axes): # diffusers/src/diffusers/models/embeddings.py:696 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['enumerate'], accessed_by=DictGetItemGuardAccessor(enumerate) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['enumerate'], 139786895625344) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__builtins_dict___4']['isinstance'], accessed_by=DictGetItemGuardAccessor(isinstance) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___4']['isinstance'], 139786873098448) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'], accessed_by=DictGetItemGuardAccessor(__import_peft_dot_tuners_dot_tuners_utils) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'], 139781156739152) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, accessed_by=GetAttrGuardAccessor(BaseTunerLayer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, 94650012074624) # from peft.tuners.tuners_utils import BaseTunerLayer # diffusers/src/diffusers/utils/peft_utils.py:113 in scale_lora_layers V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention'], 139780514175456) # if len(args) > 0 or kwargs.get("scale", None) is not None: # diffusers/src/diffusers/models/attention.py:1197 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_embeddings) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'], 139780515771264) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np, accessed_by=GetAttrGuardAccessor(np) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np, 139786867275984) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, accessed_by=GetAttrGuardAccessor(ndarray) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, 139783921183872) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math, accessed_by=GetAttrGuardAccessor(math) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math, 139786870403888) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math.log, accessed_by=GetAttrGuardAccessor(log) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math.log, 139786870406768) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch, 139786870267216) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, 139786867250080) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, accessed_by=GetAttrGuardAccessor(cos) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, 139786867251280) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, accessed_by=GetAttrGuardAccessor(exp) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, 139786867252640) # emb = torch.exp(exponent) # diffusers/src/diffusers/models/embeddings.py:62 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, accessed_by=GetAttrGuardAccessor(sin) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, 139786867261040) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, accessed_by=GetAttrGuardAccessor(outer) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, 139786867273104) # freqs = torch.outer(pos, freqs) # type: ignore # [S, D/2] # diffusers/src/diffusers/models/embeddings.py:616 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, accessed_by=GetAttrGuardAccessor(stack) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, 139786867198048) # x_rotated = torch.stack([-x_imag, x_real], dim=-1).flatten(3) # diffusers/src/diffusers/models/embeddings.py:662 in apply_rotary_emb V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, accessed_by=GetAttrGuardAccessor(arange) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, 139786867122368) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float32 == torch.float32 # start=0, end=half_dim, dtype=torch.float32, device=timesteps.device # diffusers/src/diffusers/models/embeddings.py:58 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float64, accessed_by=GetAttrGuardAccessor(float64) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float64 == torch.float64 # freqs_dtype = torch.float32 if is_mps else torch.float64 # diffusers/src/diffusers/models/embeddings.py:695 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=GetAttrGuardAccessor(apply_rotary_emb) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, 139780515561872) # from .embeddings import apply_rotary_emb # diffusers/src/diffusers/models/attention_processor.py:1760 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], 139786895605696) # if use_real: # diffusers/src/diffusers/models/embeddings.py:653 in apply_rotary_emb V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1], accessed_by=GetItemGuardAccessor(1) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1] == -1 # if use_real_unbind_dim == -1: # diffusers/src/diffusers/models/embeddings.py:659 in apply_rotary_emb V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=GetAttrGuardAccessor(get_timestep_embedding) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, 139780515482768) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3] == 10000 # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=GetAttrGuardAccessor(get_1d_rotary_pos_embed) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, 139780515495616) # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=FuncDefaultsGuardAccessor V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0] == 10000.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2], accessed_by=GetItemGuardAccessor(2) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2] == 1.0 # 1.0 # diffusers/src/diffusers/models/embeddings.py:612 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3] == 1.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'], accessed_by=DictGetItemGuardAccessor(__import_torch_dot_nn_dot_modules_dot_module) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'], 139781214054384) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch, 139786870267216) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, accessed_by=GetAttrGuardAccessor(_C) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, 139786866637808) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, accessed_by=GetAttrGuardAccessor(_get_tracing_state) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, 139781219067088) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_pre_hooks) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_activations) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'], 139780515377728) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F, 139781213067664) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F.gelu, accessed_by=GetAttrGuardAccessor(gelu) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F.gelu, 139781219449200) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_normalization) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'], 139780513909392) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch, 139786870267216) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, accessed_by=GetAttrGuardAccessor(chunk) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, 139786867250240) # scale, shift = torch.chunk(emb, 2, dim=1) # diffusers/src/diffusers/models/normalization.py:305 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, accessed_by=GetAttrGuardAccessor(rsqrt) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, 139786867196688) # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float16 == torch.float16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float32 == torch.float32 # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16, accessed_by=GetAttrGuardAccessor(bfloat16) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16 == torch.bfloat16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention_processor) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'], 139780515379008) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F, 139781213067664) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, accessed_by=GetAttrGuardAccessor(scaled_dot_product_attention) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, 139781219601200) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch, 139786870267216) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, 139786867250080) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect, accessed_by=GetAttrGuardAccessor(inspect) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect, 139786871160896) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature, accessed_by=GetAttrGuardAccessor(signature) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, 139786868738336) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:17:33.509000 139786888062784 torch/_dynamo/guards.py:2148] [0/2] [__guards] V0909 15:18:52.922000 139786888062784 torch/_dynamo/guards.py:2611] [0/3] [__recompiles] Recompiling function forward in /home/sayak/diffusers/src/diffusers/models/transformers/transformer_flux.py:388 V0909 15:18:52.922000 139786888062784 torch/_dynamo/guards.py:2611] [0/3] [__recompiles] triggered by the following guard failure(s): V0909 15:18:52.922000 139786888062784 torch/_dynamo/guards.py:2611] [0/3] [__recompiles] - len(L['self'].transformer_blocks[0].ff.net[0].proj.scaling) == 2 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:18:52.922000 139786888062784 torch/_dynamo/guards.py:2611] [0/3] [__recompiles] - len(L['self'].transformer_blocks[0].ff.net[0].proj.scaling) == 1 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:18:52.922000 139786888062784 torch/_dynamo/guards.py:2611] [0/3] [__recompiles] - ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139777664822256) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward W0909 15:21:14.853000 139786888062784 torch/fx/experimental/symbolic_shapes.py:4449] [0/3] xindex is not in var_ranges, defaulting to unknown range. V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2169] [0/3] [__guards] GUARDS: V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] TREE_GUARD_MANAGER: V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] +- RootGuardManager V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- DEFAULT_DEVICE: utils_device.CURRENT_DEVICE == None # _dynamo/output_graph.py:460 in init_ambient_guards V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GLOBAL_STATE: ___check_global_state() V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['self'], accessed_by=DictGetItemGuardAccessor(self) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- ID_MATCH: ___check_obj_id(L['self'], 139777664825136) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=L['self'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=L['self'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(L['self'].training, 139786895605728) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=L['self']._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].norm_out, accessed_by=DictGetItemGuardAccessor(norm_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out, 139777661036496) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].norm_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].norm_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.training, 139786895605728) # hidden_states = self.norm_out(hidden_states, temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:548 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm, 139777661037408) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale)[:, None, :] + shift[:, None, :] # diffusers/src/diffusers/models/normalization.py:306 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu, 139777661037312) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.silu.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].norm_out.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear, 139777661037360) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].norm_out.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].norm_out.linear.training, 139786895605728) # emb = self.linear(self.silu(conditioning_embedding).to(x.dtype)) # diffusers/src/diffusers/models/normalization.py:304 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].norm_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out, 139777661037264) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].proj_out.training, 139786895605728) # output = self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:549 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].pos_embed, accessed_by=DictGetItemGuardAccessor(pos_embed) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed, 139777664824800) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].pos_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].pos_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].pos_embed.training, 139786895605728) # image_rotary_emb = self.pos_embed(ids) # diffusers/src/diffusers/models/transformers/transformer_flux.py:469 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim, accessed_by=DictGetItemGuardAccessor(axes_dim) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].pos_embed.axes_dim, 139786895700800) # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- LENGTH_CHECK: len(L['self'].pos_embed.axes_dim) == 3 # self.axes_dim[i], pos[:, i], repeat_interleave_real=True, use_real=True, freqs_dtype=freqs_dtype # diffusers/src/diffusers/models/embeddings.py:698 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[0], accessed_by=TupleGetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[0] == 16 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[1], accessed_by=TupleGetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[1] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].pos_embed.axes_dim[2], accessed_by=TupleGetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- EQUALS_MATCH: L['self'].pos_embed.axes_dim[2] == 56 # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].pos_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].x_embedder, accessed_by=DictGetItemGuardAccessor(x_embedder) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder, 139777664824032) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].x_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].x_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].x_embedder.training, 139786895605728) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].time_text_embed, accessed_by=DictGetItemGuardAccessor(time_text_embed) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed, 139777664824848) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].time_text_embed.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.training, 139786895605728) # else self.time_text_embed(timestep, guidance, pooled_projections) # diffusers/src/diffusers/models/transformers/transformer_flux.py:452 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj, accessed_by=DictGetItemGuardAccessor(time_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj, 139777664824752) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.time_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.training, 139786895605728) # timesteps_proj = self.time_proj(timestep) # diffusers/src/diffusers/models/embeddings.py:1059 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.scale, accessed_by=DictGetItemGuardAccessor(scale) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.scale == 1 # scale=self.scale, # diffusers/src/diffusers/models/embeddings.py:769 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.num_channels, accessed_by=DictGetItemGuardAccessor(num_channels) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.num_channels == 256 # self.num_channels, # diffusers/src/diffusers/models/embeddings.py:766 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.flip_sin_to_cos, accessed_by=DictGetItemGuardAccessor(flip_sin_to_cos) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.time_proj.flip_sin_to_cos, 139786895605696) # flip_sin_to_cos=self.flip_sin_to_cos, # diffusers/src/diffusers/models/embeddings.py:767 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.time_proj.downscale_freq_shift, accessed_by=DictGetItemGuardAccessor(downscale_freq_shift) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- EQUALS_MATCH: L['self'].time_text_embed.time_proj.downscale_freq_shift == 0 # downscale_freq_shift=self.downscale_freq_shift, # diffusers/src/diffusers/models/embeddings.py:768 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder, accessed_by=DictGetItemGuardAccessor(text_embedder) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder, 139777664824416) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.text_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.training, 139786895605728) # pooled_projections = self.text_embedder(pooled_projection) # diffusers/src/diffusers/models/embeddings.py:1067 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1, accessed_by=DictGetItemGuardAccessor(act_1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1, 139777664824128) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.act_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.act_1.training, 139786895605728) # hidden_states = self.act_1(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1511 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1, 139777664824176) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_1.training, 139786895605728) # hidden_states = self.linear_1(caption) # diffusers/src/diffusers/models/embeddings.py:1510 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2, 139777664824080) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.text_embedder.linear_2.training, 139786895605728) # hidden_states = self.linear_2(hidden_states) # diffusers/src/diffusers/models/embeddings.py:1512 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.text_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder, accessed_by=DictGetItemGuardAccessor(guidance_embedder) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder, 139777664824560) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.guidance_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.training, 139786895605728) # guidance_emb = self.guidance_embedder(guidance_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1063 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1, 139777664824512) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2, 139777664824464) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.guidance_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder, accessed_by=DictGetItemGuardAccessor(timestep_embedder) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder, 139777664824704) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].time_text_embed.timestep_embedder.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.training, 139786895605728) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act, accessed_by=DictGetItemGuardAccessor(act) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act, 139780515431088) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.act.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.act.training, 139786895605728) # if self.act is not None: # diffusers/src/diffusers/models/embeddings.py:745 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1, accessed_by=DictGetItemGuardAccessor(linear_1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1, 139777664824656) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_1.training, 139786895605728) # sample = self.linear_1(sample) # diffusers/src/diffusers/models/embeddings.py:743 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2, accessed_by=DictGetItemGuardAccessor(linear_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2, 139777664824608) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.linear_2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.linear_2.training, 139786895605728) # sample = self.linear_2(sample) # diffusers/src/diffusers/models/embeddings.py:748 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.post_act, accessed_by=DictGetItemGuardAccessor(post_act) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.post_act, 139786895690912) # if self.post_act is not None: # diffusers/src/diffusers/models/embeddings.py:750 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].time_text_embed.timestep_embedder.forward.__defaults__[0], 139786895690912) # if condition is not None: # diffusers/src/diffusers/models/embeddings.py:741 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].time_text_embed._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].context_embedder, accessed_by=DictGetItemGuardAccessor(context_embedder) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder, 139777664824320) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].context_embedder.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].context_embedder.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].context_embedder.training, 139786895605728) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].transformer_blocks, accessed_by=DictGetItemGuardAccessor(transformer_blocks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks, 139777664823936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0], 139777664823984) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff, 139777664822544) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net, 139777664822304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0], 139777664822352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj, 139728980164176) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A, 139728980157744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'], 139727515387648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_A['default_2'].weight, 139726874038720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B, 139728980152176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['default_2'], 139727515390288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.base_layer, 139777664822256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout, 139728980150064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['default_2'], 139727515392928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1], 139777664822208) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2], 139728980164416) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A, 139728979916928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'], 139727515390096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_A['default_2'].weight, 139726874038480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B, 139728979915584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B['default_2'], 139727515391248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].base_layer, 139777664822160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout, 139728979906896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout['default_2'], 139727515389904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn, 139777664823360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k, 139728979952400) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A, 139728979958400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'], 139727515396672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_A['default_2'].weight, 139726876414080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B, 139728979965600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B['default_2'], 139727515396384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.base_layer, 139777664823168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout, 139728979953168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout['default_2'], 139727515387360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q, 139728980350656) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A, 139728980357280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'], 139727515397488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_A['default_2'].weight, 139726876413600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B, 139728980353008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B['default_2'], 139727515392688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.base_layer, 139777664823072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout, 139728980358864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout['default_2'], 139727515397584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v, 139728979954416) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A, 139728979960560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'], 139727515400752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_A['default_2'].weight, 139726876414640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B, 139728979958352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B['default_2'], 139727515397728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.base_layer, 139777664822976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout, 139728979964784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout['default_2'], 139727515400320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k, 139777664823120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_k.weight, 139777659850240) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q, 139777664823216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_q.weight, 139777660412416) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out, 139777664822784) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0], 139728980312704) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A, 139728980320144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'], 139727515390960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_A['default_2'].weight, 139726874047600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B, 139728980322880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B['default_2'], 139727515392400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].base_layer, 139777664822736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout, 139728980320048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['default_2'], 139727515402192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1], 139777664822688) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj, 139728980321920) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A, 139728980321872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'], 139727515402096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_A['default_2'].weight, 139726876415200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B, 139728980322112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['default_2'], 139727515400800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.base_layer, 139777664822928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout, 139728980321680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['default_2'], 139727515401712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj, 139728980317456) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A, 139728980318704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'], 139727515393504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_A['default_2'].weight, 139726874049200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B, 139728980328400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['default_2'], 139727515386112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.base_layer, 139777664822832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout, 139728980318800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['default_2'], 139727515386160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj, 139728980327920) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A, 139728980320768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'], 139727515399168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_A['default_2'].weight, 139726876415680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B, 139728980317024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['default_2'], 139727515401376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.base_layer, 139777664822880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout, 139728980316016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['default_2'], 139727515395856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out, 139728978514016) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A, 139728978525632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'], 139727515398784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_A['default_2'].weight, 139726874048800) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B, 139728980148576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B['default_2'], 139727515394272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.base_layer, 139777664822640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout, 139728978523904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['default_2'], 139727515398832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k, 139777664822496) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_k.weight, 139777664182976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q, 139777664822592) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.norm_added_q.weight, 139777664182016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].attn.processor, 139777664823408) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1, 139777664823840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm, 139777664823648) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu, 139777664823744) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear, 139728989471024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A, 139728989245152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'], 139727515392544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_A['default_2'].weight, 139726886206368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B, 139728966627440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B['default_2'], 139727515398016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.base_layer, 139777664823696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout, 139728989244624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout['default_2'], 139727515399936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear._active_adapter, 139786895671040) # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].norm1.linear._active_adapter) == 1 # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._active_adapter[0], accessed_by=ListGetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1.linear._active_adapter[0] == 'default_2' # for active_adapter in self.active_adapters: # peft/tuners/lora/layer.py:559 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2, 139777664822448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context, 139777664822112) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net, 139777664821968) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[0].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0], 139777664822016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj, 139728979908240) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A, 139728979911984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'], 139727515387984) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_A['default_2'].weight, 139726874043120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B, 139728979908624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['default_2'], 139727515387120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer, 139777664821920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout, 139728979903488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['default_2'], 139727515387552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1], 139777664821824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2], 139728978644320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A, 139728978645328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'], 139727515400272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_A['default_2'].weight, 139726874045920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B, 139728978644800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B['default_2'], 139727515401616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].base_layer, 139777664821776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout, 139728978644176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['default_2'], 139727515399312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context, 139777664823600) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm, 139777664823456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu, 139777664823552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear, 139728978846736) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[0].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A, 139728978842752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'], 139727515400224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_A['default_2'].weight, 139726871114784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B, 139728978844096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B['default_2'], 139727515399216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.base_layer, 139777664823504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout, 139728978850096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['default_2'], 139727515399888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[0].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[0].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[0].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[0].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context, 139777664822400) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[0].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1], 139777664823888) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff, 139777664820480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net, 139777664820240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0], 139777664820288) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj, 139728979049056) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A, 139728979048768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'], 139727515400656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_A['default_2'].weight, 139726872848928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B, 139728979036480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['default_2'], 139727515399504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.base_layer, 139777664820192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout, 139728979036816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['default_2'], 139727515398976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1], 139777664820144) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2], 139728978900976) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A, 139728978901648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'], 139727515396096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_A['default_2'].weight, 139726872848048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B, 139728978902032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B['default_2'], 139727515394464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].base_layer, 139777664820096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout, 139728978892768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout['default_2'], 139727515396048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn, 139777664821248) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k, 139728978656416) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A, 139728978970128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'], 139727515397296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_A['default_2'].weight, 139726872852688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B, 139728978969408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B['default_2'], 139727515397392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.base_layer, 139777664821104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout, 139728978970272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout['default_2'], 139727515397344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q, 139728978650272) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A, 139728978649456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'], 139727515401904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_A['default_2'].weight, 139726872853248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B, 139728978643168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B['default_2'], 139727515399648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.base_layer, 139777664821008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout, 139728978650368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout['default_2'], 139727515401088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v, 139728978974208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A, 139728978972288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'], 139727515387024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_A['default_2'].weight, 139726872852368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B, 139728978972192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B['default_2'], 139727515388320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.base_layer, 139777664820912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout, 139728978975792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout['default_2'], 139727515388464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k, 139777664821056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_k.weight, 139777660113664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q, 139777664821152) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_q.weight, 139777659965328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out, 139777664820720) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0], 139728980484416) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A, 139728980488928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'], 139727515389136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_A['default_2'].weight, 139726872849808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B, 139728980483696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B['default_2'], 139727515389664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].base_layer, 139777664820672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout, 139728980488496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['default_2'], 139727515389040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1], 139777664820624) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj, 139728978977424) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A, 139728978798256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'], 139727515390240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_A['default_2'].weight, 139726872851728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B, 139728978795856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['default_2'], 139727515400416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.base_layer, 139777664820864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout, 139728978974592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['default_2'], 139727515390192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj, 139728980442464) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A, 139728980489888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'], 139727515393792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_A['default_2'].weight, 139726872849968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B, 139728980482064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['default_2'], 139727515396912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.base_layer, 139777664820768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout, 139728980490128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['default_2'], 139727515393744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj, 139728989471600) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A, 139728980438096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'], 139727515388752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_A['default_2'].weight, 139726872850848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B, 139728980442896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['default_2'], 139727515398496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.base_layer, 139777664820816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout, 139728980438720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['default_2'], 139727515392160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out, 139728980489408) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A, 139728980490896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'], 139727515396576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_A['default_2'].weight, 139726872849088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B, 139728980476928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B['default_2'], 139727515389568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.base_layer, 139777664820576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout, 139728980491664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['default_2'], 139727515396528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k, 139777664820432) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_k.weight, 139777659965408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q, 139777664820528) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.norm_added_q.weight, 139777659965488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].attn.processor, 139777664821296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1, 139777664821728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm, 139777664821584) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu, 139777664821680) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear, 139728978654064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A, 139728978653872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'], 139727515391152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_A['default_2'].weight, 139726872854208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B, 139728978652672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B['default_2'], 139727515395904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.base_layer, 139777664821632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout, 139728978654160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout['default_2'], 139727515391728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2, 139777664820384) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context, 139777664820048) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net, 139777664819904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[1].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0], 139777664819952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj, 139728978901744) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A, 139728978899200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'], 139727515387792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_A['default_2'].weight, 139726872847488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B, 139728978900880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['default_2'], 139727515396624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer, 139777664819856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout, 139728978902416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['default_2'], 139727515388896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1], 139777664819760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2], 139728978888448) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A, 139728978898384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'], 139727515385968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_A['default_2'].weight, 139726872846768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B, 139728978900064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B['default_2'], 139727515395040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].base_layer, 139777664819712) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout, 139728978901264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['default_2'], 139727515394896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context, 139777664821536) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm, 139777664821344) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu, 139777664821440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear, 139728978647152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[1].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A, 139728978647008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'], 139727515395520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_A['default_2'].weight, 139726872853648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B, 139728978644464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B['default_2'], 139727515395136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.base_layer, 139777664821392) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout, 139728978646624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['default_2'], 139727515395472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[1].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[1].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[1].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[1].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context, 139777664820336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[1].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2], 139777664822064) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff, 139777664818416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net, 139777664818176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0], 139777664818224) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj, 139728980519488) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A, 139728980519392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'], 139727513728624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_A['default_2'].weight, 139726872842688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B, 139728980519872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['default_2'], 139727513719840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.base_layer, 139777664818128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout, 139728980519536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['default_2'], 139727513727808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1], 139777664818080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2], 139728980516320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A, 139728980524576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'], 139727513720464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_A['default_2'].weight, 139726872842928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B, 139728980522800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B['default_2'], 139727513720320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].base_layer, 139777664818032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout, 139728980517472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout['default_2'], 139727513720224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn, 139777664819184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k, 139728977639600) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A, 139728977639744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'], 139727513720368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_A['default_2'].weight, 139726872845088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B, 139728977639696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B['default_2'], 139727513725600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.base_layer, 139777664819040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout, 139728977638784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout['default_2'], 139727513726320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q, 139728978789040) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A, 139728977625488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'], 139727513725264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_A['default_2'].weight, 139726872845568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B, 139728977625152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B['default_2'], 139727513725408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.base_layer, 139777664818944) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout, 139728977632256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout['default_2'], 139727513725312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v, 139728977632784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A, 139728977637488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'], 139727513723488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_A['default_2'].weight, 139726872850208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B, 139728977626256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B['default_2'], 139727513725696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.base_layer, 139777664818848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout, 139728977634896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout['default_2'], 139727513726272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k, 139777664818992) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_k.weight, 139777664567632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q, 139777664819088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_q.weight, 139777664569792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out, 139777664818656) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0], 139728979413168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A, 139728979426464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'], 139727513722144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_A['default_2'].weight, 139726872840928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B, 139728979426368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B['default_2'], 139727513726992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].base_layer, 139777664818608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout, 139728979422144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['default_2'], 139727513721952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1], 139777664818560) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj, 139728978452848) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A, 139728978454480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'], 139727513722192) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_A['default_2'].weight, 139726872840768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B, 139728978456976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['default_2'], 139727513725840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.base_layer, 139777664818800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout, 139728978444448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['default_2'], 139727513727616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj, 139728978506368) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A, 139728978499120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'], 139727513725552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_A['default_2'].weight, 139726872841408) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B, 139728978499360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['default_2'], 139727513724784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.base_layer, 139777664818704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout, 139728978500992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['default_2'], 139727513720800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj, 139728978458032) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A, 139728978458608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'], 139727513719408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_A['default_2'].weight, 139726872840048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B, 139728978456064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['default_2'], 139727513725936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.base_layer, 139777664818752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout, 139728978450208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['default_2'], 139727513720080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out, 139728979423344) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A, 139728979419120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'], 139727513716912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_A['default_2'].weight, 139726872841808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B, 139728979420656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B['default_2'], 139727513716336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.base_layer, 139777664818512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout, 139728979421040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['default_2'], 139727513728192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k, 139777664818368) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_k.weight, 139777664568672) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q, 139777664818464) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.norm_added_q.weight, 139777664567552) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].attn.processor, 139777664819232) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1, 139777664819664) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm, 139777664819520) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu, 139777664819616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear, 139728978887296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A, 139728978887584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'], 139727513729056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_A['default_2'].weight, 139726872846448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B, 139728978887824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B['default_2'], 139727513716384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.base_layer, 139777664819568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout, 139728978897568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout['default_2'], 139727515396864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2, 139777664818320) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context, 139777664817984) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net, 139777664817840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[2].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0], 139777664817888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj, 139728980516272) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A, 139728980512000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'], 139727513727328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_A['default_2'].weight, 139726872843488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B, 139728979094896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['default_2'], 139727513726752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer, 139777664817792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout, 139728980522944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['default_2'], 139727513727424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1], 139777664817696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2], 139728979089472) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A, 139728979092208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'], 139727513730544) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_A['default_2'].weight, 139726872843648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B, 139728979097440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B['default_2'], 139727513722912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].base_layer, 139777664817648) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout, 139728979097632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['default_2'], 139727513730592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context, 139777664819472) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm, 139777664819280) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu, 139777664819376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear, 139728978901024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[2].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A, 139728978900784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'], 139727513724208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_A['default_2'].weight, 139726872846048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B, 139728978900256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B['default_2'], 139727513726512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.base_layer, 139777664819328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout, 139728978888160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['default_2'], 139727513728288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[2].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[2].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[2].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[2].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context, 139777664818272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[2].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3], 139777664820000) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff, 139777664816352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net, 139777664816112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0], 139777664816160) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj, 139728977664448) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A, 139728977662576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'], 139727513964352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_A['default_2'].weight, 139726880999680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B, 139728977663920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['default_2'], 139727513969632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.base_layer, 139777664816064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout, 139728977665504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['default_2'], 139727513969728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1], 139777664816016) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2], 139728977672992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A, 139728977665408) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'], 139727513971840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_A['default_2'].weight, 139726881003120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B, 139728977664592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B['default_2'], 139727513970880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].base_layer, 139777664815968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout, 139728977665168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout['default_2'], 139727513971648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn, 139777664817120) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k, 139728979281472) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A, 139728979293040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'], 139727513731024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_A['default_2'].weight, 139726872839088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B, 139728979293520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B['default_2'], 139727513724544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.base_layer, 139777664816976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout, 139728979280608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout['default_2'], 139727513724400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q, 139728979284784) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A, 139728979283584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'], 139727513725120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_A['default_2'].weight, 139726872838528) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B, 139728979283536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B['default_2'], 139727513724592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.base_layer, 139777664816880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout, 139728979286608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout['default_2'], 139727513725216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v, 139728979144432) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A, 139728979137280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'], 139727513720944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_A['default_2'].weight, 139726872839488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B, 139728979137616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B['default_2'], 139727513721520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.base_layer, 139777664816784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout, 139728979136608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout['default_2'], 139727513726032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k, 139777664816928) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_k.weight, 139777659970608) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q, 139777664817024) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_q.weight, 139777664565792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out, 139777664816592) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0], 139728978811136) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A, 139728978811184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'], 139727513968720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_A['default_2'].weight, 139727352597216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B, 139728978817136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B['default_2'], 139727513968624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].base_layer, 139777664816544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout, 139728978820496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['default_2'], 139727513969776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1], 139777664816496) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj, 139728979138432) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A, 139728979135456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'], 139727513721040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_A['default_2'].weight, 139726886604448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B, 139728979138480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['default_2'], 139727513720656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.base_layer, 139777664816736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout, 139728979135216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['default_2'], 139727513720848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj, 139728979142224) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A, 139728978807968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'], 139727513965888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_A['default_2'].weight, 139727352313264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B, 139728978820160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['default_2'], 139727513965024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.base_layer, 139777664816640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout, 139728978819824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['default_2'], 139727513966416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj, 139728979142320) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A, 139728979133200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'], 139727513961328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_A['default_2'].weight, 139727271006976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B, 139728979136656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['default_2'], 139727513961760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.base_layer, 139777664816688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout, 139728979139776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['default_2'], 139727513961280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out, 139728978810896) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A, 139728978815984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'], 139727513969440) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_A['default_2'].weight, 139726890394672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B, 139728978816224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B['default_2'], 139727513969248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.base_layer, 139777664816448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout, 139728978819920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['default_2'], 139727513969296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k, 139777664816304) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_k.weight, 139777659970448) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q, 139777664816400) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.norm_added_q.weight, 139777659970528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].attn.processor, 139777664817168) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1, 139777664817600) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm, 139777664817456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu, 139777664817552) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear, 139728978751648) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A, 139728978754192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'], 139727513727664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_A['default_2'].weight, 139726872844448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B, 139728978746656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B['default_2'], 139727513716816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.base_layer, 139777664817504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout, 139728978751696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout['default_2'], 139727513727520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2, 139777664816256) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context, 139777664815920) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net, 139777664815776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[3].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0], 139777664815824) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj, 139728979338592) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A, 139728979342672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'], 139727513972608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_A['default_2'].weight, 139726881010800) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B, 139728979342000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['default_2'], 139727513971936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer, 139777664815728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout, 139728979339600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['default_2'], 139727513972368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1], 139777664815632) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2], 139728977447568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A, 139728979210256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'], 139727513973904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_A['default_2'].weight, 139726881000320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B, 139728979212656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B['default_2'], 139727513973136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].base_layer, 139777664815584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout, 139728977456496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['default_2'], 139727513973856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context, 139777664817408) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm, 139777664817216) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu, 139777664817312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear, 139728979286992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[3].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A, 139728979286800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'], 139727513718640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_A['default_2'].weight, 139726872841568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B, 139728979284832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B['default_2'], 139727513720176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.base_layer, 139777664817264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout, 139728979287088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['default_2'], 139727513725888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[3].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[3].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[3].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[3].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context, 139777664816208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[3].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4], 139777664817936) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff, 139777664814288) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net, 139777664814048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0], 139777664814096) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj, 139728977554704) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A, 139728977558832) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'], 139727513962624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_A['default_2'].weight, 139726881012000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B, 139728978036816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['default_2'], 139727513974912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.base_layer, 139777664814000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout, 139728977558592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['default_2'], 139727513963824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1], 139777664813952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2], 139728978048528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A, 139728978046800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'], 139727513968912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_A['default_2'].weight, 139726881012400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B, 139728978035664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B['default_2'], 139727513961520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].base_layer, 139777664813904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout, 139728978037584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout['default_2'], 139727513968864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn, 139777664815056) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k, 139728978900544) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A, 139728977775632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'], 139727513970352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_A['default_2'].weight, 139726880998960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B, 139728977773328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B['default_2'], 139727513976592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.base_layer, 139777664814912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout, 139728977775248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout['default_2'], 139727513975152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q, 139728979210304) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A, 139728979375248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'], 139727513976400) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_A['default_2'].weight, 139726881001760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B, 139728979362960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B['default_2'], 139727513975392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.base_layer, 139777664814816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout, 139728979373088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout['default_2'], 139727513974720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v, 139728977774192) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A, 139728977773136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'], 139727513970496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_A['default_2'].weight, 139726881006720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B, 139728977773184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B['default_2'], 139727513971216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.base_layer, 139777664814720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout, 139728977774528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout['default_2'], 139727513970784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k, 139777664814864) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_k.weight, 139777664570512) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q, 139777664814960) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_q.weight, 139777664571712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out, 139777664814528) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0], 139728979356560) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A, 139728979356416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'], 139727513964112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_A['default_2'].weight, 139726881008480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B, 139728979361408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B['default_2'], 139727513963248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].base_layer, 139777664814480) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout, 139728979348784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['default_2'], 139727513964016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1], 139777664814432) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj, 139728977779760) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A, 139728977788064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'], 139727513961040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_A['default_2'].weight, 139726881007120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B, 139728977780576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['default_2'], 139727513963056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.base_layer, 139777664814672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout, 139728977788784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['default_2'], 139727513963344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj, 139728980379200) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A, 139728980390816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'], 139727513964736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_A['default_2'].weight, 139726881008080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B, 139728980380976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['default_2'], 139727513964880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.base_layer, 139777664814576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout, 139728980381888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['default_2'], 139727513965216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj, 139728977773808) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A, 139728980384576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'], 139727513966704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_A['default_2'].weight, 139726881007680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B, 139728980389328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['default_2'], 139727513966224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.base_layer, 139777664814624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout, 139728980385968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['default_2'], 139727513966320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out, 139728977554176) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A, 139728977557488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'], 139727513967712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_A['default_2'].weight, 139726881011520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B, 139728977556720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B['default_2'], 139727513967760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.base_layer, 139777664814384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout, 139728977549616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['default_2'], 139727513967664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k, 139777664814240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_k.weight, 139777659971408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q, 139777664814336) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.norm_added_q.weight, 139777664568832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].attn.processor, 139777664815104) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1, 139777664815536) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm, 139777664815392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu, 139777664815488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear, 139728979208528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A, 139728979203584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'], 139727513971120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_A['default_2'].weight, 139726881000720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B, 139728979206608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B['default_2'], 139727513969824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.base_layer, 139777664815440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout, 139728979206416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout['default_2'], 139727513970832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2, 139777664814192) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context, 139777664813856) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net, 139777664813712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[4].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0], 139777664813760) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj, 139728979365408) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A, 139728976457936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'], 139727513967952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_A['default_2'].weight, 139726881012880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B, 139728976456784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['default_2'], 139727513966128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer, 139777664813664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout, 139728976447712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['default_2'], 139727513968048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1], 139777664813568) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2], 139728976450976) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A, 139728976460432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'], 139727513967040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_A['default_2'].weight, 139726881006400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B, 139728976451360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B['default_2'], 139727513967328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].base_layer, 139777664813520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout, 139728976451264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['default_2'], 139727513967472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context, 139777664815344) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm, 139777664815152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu, 139777664815248) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear, 139728979209344) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[4].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A, 139728979207856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'], 139727513974816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_A['default_2'].weight, 139726881001200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B, 139728979212560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B['default_2'], 139727513974624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.base_layer, 139777664815200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout, 139728979203008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['default_2'], 139727513974384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[4].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[4].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[4].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[4].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context, 139777664814144) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[4].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5], 139777664815872) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff, 139777664812224) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net, 139777664811984) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0], 139777664812032) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj, 139728978177056) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A, 139728979063520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'], 139727514193056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_A['default_2'].weight, 139726880211728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B, 139728979062752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['default_2'], 139727514205488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.base_layer, 139777664811936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout, 139728979054592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['default_2'], 139727514205968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1], 139777664811888) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2], 139728979052672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A, 139728979058912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'], 139727514192960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_A['default_2'].weight, 139726880223008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B, 139728978307696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B['default_2'], 139727514193488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].base_layer, 139777664811840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout, 139728979052960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout['default_2'], 139727514193440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn, 139777664812992) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k, 139728977932304) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A, 139728977935808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'], 139727513971168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_A['default_2'].weight, 139726880998400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B, 139728977934464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B['default_2'], 139727513973568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.base_layer, 139777664812848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout, 139728977923232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout['default_2'], 139727513970736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q, 139728976378944) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A, 139728976373232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'], 139727513975056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_A['default_2'].weight, 139726880997760) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B, 139728976377984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B['default_2'], 139727513962336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.base_layer, 139777664812752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout, 139728976377840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout['default_2'], 139727513976016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v, 139728977920784) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A, 139728977926160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'], 139727513968240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_A['default_2'].weight, 139726881003520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B, 139728978168224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B['default_2'], 139727513971984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.base_layer, 139777664812656) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout, 139728977931776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout['default_2'], 139727513961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k, 139777664812800) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_k.weight, 139781198830368) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q, 139777664812896) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_q.weight, 139777664567072) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out, 139777664812464) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0], 139728978179600) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A, 139728978178160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'], 139727514195840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_A['default_2'].weight, 139726881002000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B, 139728978176816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B['default_2'], 139727514198864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].base_layer, 139777664812416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout, 139728978173360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['default_2'], 139727514198912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1], 139777664812368) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj, 139728978181328) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A, 139728978181280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'], 139727513966512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_A['default_2'].weight, 139726881009040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B, 139728978180512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['default_2'], 139727513968192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.base_layer, 139777664812608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout, 139728978181376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['default_2'], 139727513966560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj, 139728978176864) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A, 139728978176480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'], 139727514197136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_A['default_2'].weight, 139726881010960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B, 139728978176048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['default_2'], 139727514197712) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.base_layer, 139777664812512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout, 139728978173984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['default_2'], 139727513972080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj, 139728978175808) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A, 139728978174080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'], 139727513965792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_A['default_2'].weight, 139726881009520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B, 139728978169760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['default_2'], 139727513967088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.base_layer, 139777664812560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout, 139728978180032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['default_2'], 139727513965504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out, 139728978171728) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A, 139728978171440) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'], 139727514204288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_A['default_2'].weight, 139726881010000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B, 139728978173648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B['default_2'], 139727514203184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.base_layer, 139777664812320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout, 139728978178016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['default_2'], 139727514204192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k, 139777664812176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_k.weight, 139781198834848) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q, 139777664812272) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.norm_added_q.weight, 139777664184096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].attn.processor, 139777664813040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1, 139777664813472) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm, 139777664813328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu, 139777664813424) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear, 139728977894256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A, 139728976365840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'], 139727513970976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_A['default_2'].weight, 139726881000000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B, 139728976367952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B['default_2'], 139727513971312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.base_layer, 139777664813376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout, 139728977894640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout['default_2'], 139727513973280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2, 139777664812128) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context, 139777664811792) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net, 139777664811648) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[5].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0], 139777664811696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj, 139728978312400) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A, 139728978302128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'], 139727514200880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_A['default_2'].weight, 139726880220768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B, 139728978306352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['default_2'], 139727514193776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer, 139777664811600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout, 139728978307552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['default_2'], 139727514193824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1], 139777664811504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2], 139728976484848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A, 139728976478656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'], 139727514191520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_A['default_2'].weight, 139726880221728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B, 139728976488352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B['default_2'], 139727514200928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].base_layer, 139777664811456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout, 139728976492384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['default_2'], 139727514200640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context, 139777664813280) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm, 139777664813088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu, 139777664813184) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear, 139728976368000) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[5].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A, 139728976366464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'], 139727513962864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_A['default_2'].weight, 139726880997440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B, 139728976366368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B['default_2'], 139727513969008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.base_layer, 139777664813136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout, 139728976367712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['default_2'], 139727513962816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[5].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[5].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[5].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[5].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context, 139777664812080) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[5].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6], 139777664813808) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff, 139777674394928) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net, 139777674394688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0], 139777674394736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj, 139728978388992) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A, 139728978390240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'], 139727514191136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_A['default_2'].weight, 139726880960288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B, 139728978390144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['default_2'], 139727514195072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.base_layer, 139777674394640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout, 139728978389664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['default_2'], 139727514190128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1], 139777674394592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2], 139728978395040) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A, 139728978381456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'], 139727514196896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_A['default_2'].weight, 139726880962208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B, 139728978385392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B['default_2'], 139727514197472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].base_layer, 139777674394544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout, 139728978392016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout['default_2'], 139727514196800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn, 139777674383792) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k, 139728977441888) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A, 139728977436224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'], 139727514195744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_A['default_2'].weight, 139726880950288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B, 139728977434160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B['default_2'], 139727514200400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.base_layer, 139777674395552) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout, 139728977443136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout['default_2'], 139727514199728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q, 139728978068288) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A, 139728978074768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'], 139727514198528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_A['default_2'].weight, 139726893206480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B, 139728978074912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B['default_2'], 139727514197040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.base_layer, 139777674395456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout, 139728978068624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout['default_2'], 139727514198768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v, 139728977959648) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A, 139728977958592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'], 139727514203328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_A['default_2'].weight, 139726880950768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B, 139728977959888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B['default_2'], 139727514199872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.base_layer, 139777674395360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout, 139728977964688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout['default_2'], 139727514195600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k, 139777674395504) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_k.weight, 139777664177376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q, 139777674395600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_q.weight, 139777659972208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out, 139777674395168) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0], 139728977486000) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A, 139728977481536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'], 139727514198144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_A['default_2'].weight, 139726880958848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B, 139728978393744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B['default_2'], 139727514198192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].base_layer, 139777674395120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout, 139728977490512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['default_2'], 139727514198096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1], 139777674395072) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj, 139728977968144) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A, 139728977825792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'], 139727514205920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_A['default_2'].weight, 139726880951248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B, 139728977834000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['default_2'], 139727514203904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.base_layer, 139777674395312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout, 139728977960032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['default_2'], 139727514205872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj, 139728977485376) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A, 139728977485232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'], 139727514192576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_A['default_2'].weight, 139726880958288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B, 139728977484944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['default_2'], 139727514193200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.base_layer, 139777674395216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout, 139728977491136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['default_2'], 139727514192624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj, 139728977834288) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A, 139728977478512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'], 139727514194208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_A['default_2'].weight, 139726880951648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B, 139728977488304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['default_2'], 139727514199248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.base_layer, 139777674395264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout, 139728977832896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['default_2'], 139727514194112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out, 139728978385440) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A, 139728978378960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'], 139727514204768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_A['default_2'].weight, 139726880959408) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B, 139728978381744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B['default_2'], 139727514191904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.base_layer, 139777674395024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout, 139728978385056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['default_2'], 139727514192192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k, 139777674394880) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_k.weight, 139777664570752) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q, 139777674394976) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.norm_added_q.weight, 139777664177456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].attn.processor, 139777674380624) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1, 139777664811408) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm, 139777664811264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu, 139777664811360) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear, 139728978080576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A, 139728978067760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'], 139727514190800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_A['default_2'].weight, 139726880224848) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B, 139728978075152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B['default_2'], 139727514192000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.base_layer, 139777664811312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout, 139728978068144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout['default_2'], 139727514204240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2, 139777674394832) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context, 139777674394496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net, 139777674394352) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[6].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0], 139777674394400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj, 139728978220448) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A, 139728978216320) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'], 139727514190752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_A['default_2'].weight, 139726880956368) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B, 139728978215744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['default_2'], 139727514194016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer, 139777674394256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout, 139728978221168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['default_2'], 139727514192048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1], 139777665008928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2], 139728978230048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A, 139728978224336) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'], 139727514192480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_A['default_2'].weight, 139726880961888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B, 139728978025664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B['default_2'], 139727514191568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].base_layer, 139777665009312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout, 139728978230576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['default_2'], 139727514192672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context, 139777664811216) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm, 139777674384080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu, 139777664811120) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear, 139728978079040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[6].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A, 139728978078752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'], 139727514204336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_A['default_2'].weight, 139726880213728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B, 139728978083504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B['default_2'], 139727514199536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.base_layer, 139777664811072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout, 139728978067616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['default_2'], 139727514197328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[6].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[6].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[6].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[6].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context, 139777674394784) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[6].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7], 139777664811744) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff, 139777663769216) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net, 139777663770080) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0], 139777663770656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj, 139728975703024) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A, 139728975697600) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'], 139722357221232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_A['default_2'].weight, 139726883372736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B, 139728975707776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['default_2'], 139722357225456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.base_layer, 139777663770416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout, 139728975695680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['default_2'], 139722357225648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1], 139777663769504) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2], 139728975700336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A, 139728975701632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'], 139722357226464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_A['default_2'].weight, 139726883370096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B, 139728975702064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B['default_2'], 139722357212592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].base_layer, 139777663769648) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout, 139728975693472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout['default_2'], 139722357226512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn, 139777665009504) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k, 139728977295104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A, 139728976543904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'], 139727514202944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_A['default_2'].weight, 139726886400176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B, 139728976546928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B['default_2'], 139727514191808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.base_layer, 139780515283440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout, 139728977297264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout['default_2'], 139727514196992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q, 139728978033104) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A, 139728978025328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'], 139727514202272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_A['default_2'].weight, 139726886402096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B, 139728978026288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B['default_2'], 139727514205392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.base_layer, 139786872078544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout, 139728978031472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout['default_2'], 139727514202224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v, 139728976546496) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A, 139728976556912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'], 139727514205104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_A['default_2'].weight, 139726886403296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B, 139728976551536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B['default_2'], 139727514205248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.base_layer, 139780515283008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout, 139728976550192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout['default_2'], 139727514205200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k, 139786872288784) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_k.weight, 139777664175376) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q, 139777676533120) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_q.weight, 139781198836928) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out, 139777663769456) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0], 139728975827376) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A, 139728975834624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'], 139722357213552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_A['default_2'].weight, 139726886401136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B, 139728975835248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B['default_2'], 139722357224496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].base_layer, 139777663769408) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout, 139728975830448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['default_2'], 139722357221616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1], 139777663769360) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj, 139728975832032) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A, 139728975831552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'], 139727514194496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_A['default_2'].weight, 139726886389616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B, 139728975830928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['default_2'], 139727514191232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.base_layer, 139780515287040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout, 139728975830880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['default_2'], 139727514194544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj, 139728975835632) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A, 139728975835152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'], 139727514200208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_A['default_2'].weight, 139726886388496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B, 139728975833088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['default_2'], 139727514204672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.base_layer, 139780515287088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout, 139728975834144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['default_2'], 139727514204960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj, 139728975836496) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A, 139728975836448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'], 139727514192816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_A['default_2'].weight, 139726886390016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B, 139728975833856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['default_2'], 139727514194736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.base_layer, 139780515287136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout, 139728975831696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['default_2'], 139727514194880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out, 139728975831168) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A, 139728975704368) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'], 139722357220656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_A['default_2'].weight, 139726886402576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B, 139728975707680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B['default_2'], 139722357218880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.base_layer, 139777663769312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout, 139728975823824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['default_2'], 139722357220944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k, 139777663769120) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_k.weight, 139777660418256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q, 139777663769264) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.norm_added_q.weight, 139777664178256) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].attn.processor, 139777665009072) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1, 139777665009360) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm, 139777665009456) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu, 139777665009408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear, 139728978027824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A, 139728978034640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'], 139727514195552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_A['default_2'].weight, 139726880962288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B, 139728978026576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B['default_2'], 139727514205584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.base_layer, 139777665008976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout, 139728978025904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout['default_2'], 139727514194928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2, 139777663769552) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context, 139777663769744) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net, 139777663769600) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[7].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0], 139777663769696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj, 139728975693808) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A, 139728975698320) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'], 139722357216816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_A['default_2'].weight, 139726883357056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B, 139728975706336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['default_2'], 139722357224064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer, 139777663770176) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout, 139728975704656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['default_2'], 139722357224736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1], 139777663770128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2], 139728976639792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A, 139728976636240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'], 139722354730336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_A['default_2'].weight, 139726883365056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B, 139728976641472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B['default_2'], 139722354731056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].base_layer, 139777663770320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout, 139728976639360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['default_2'], 139722357223344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context, 139777665009120) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm, 139777665007728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu, 139777665009024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear, 139728978029936) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[7].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A, 139728978034064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'], 139727514203376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_A['default_2'].weight, 139726886402256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B, 139728978020000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B['default_2'], 139727514203472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.base_layer, 139777665007824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout, 139728978033776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['default_2'], 139727514203424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[7].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[7].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[7].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[7].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context, 139777663770752) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[7].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8], 139777674394448) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff, 139777663771472) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net, 139777663771712) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0], 139777663771664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj, 139728975371312) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A, 139728975368768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'], 139722354721888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_A['default_2'].weight, 139726883364496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B, 139728975378800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['default_2'], 139722354721936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.base_layer, 139777663771760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout, 139728975374528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['default_2'], 139722354722176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1], 139777663771808) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2], 139728976625728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A, 139728975415328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'], 139722354729904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_A['default_2'].weight, 139726883363936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B, 139728975415040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B['default_2'], 139722354730624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].base_layer, 139777663771856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout, 139728975421328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout['default_2'], 139722354729952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn, 139777663770848) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k, 139728976642000) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A, 139728976637968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'], 139722354721072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_A['default_2'].weight, 139726883372496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B, 139728976636912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B['default_2'], 139722354720832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.base_layer, 139777663770800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout, 139728976627984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout['default_2'], 139722354720976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q, 139728976632832) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A, 139728976625872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'], 139722354720928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_A['default_2'].weight, 139726883363856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B, 139728976630720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B['default_2'], 139722354729040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.base_layer, 139777663770944) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout, 139728976637296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout['default_2'], 139722354728992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v, 139728978038448) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A, 139728975328112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'], 139722354726832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_A['default_2'].weight, 139726883371856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B, 139728975328256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B['default_2'], 139722354726592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.base_layer, 139777663771040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout, 139728975329216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout['default_2'], 139722354726640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k, 139777663770896) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_k.weight, 139777659939520) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q, 139777663770704) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_q.weight, 139777664578912) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out, 139777663771232) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0], 139728975372176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A, 139728975378704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'], 139722354724048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_A['default_2'].weight, 139726883366816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B, 139728975369728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B['default_2'], 139722354724768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].base_layer, 139777663771280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout, 139728975376160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['default_2'], 139722354723952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1], 139777663771328) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj, 139728975325568) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A, 139728976724656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'], 139722354725968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_A['default_2'].weight, 139726883357936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B, 139728976729312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['default_2'], 139722354726112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.base_layer, 139777663771088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout, 139728975325904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['default_2'], 139722354726160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj, 139728976726048) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A, 139728976737520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'], 139722354727072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_A['default_2'].weight, 139726883372576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B, 139728976737472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['default_2'], 139722354722272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.base_layer, 139777663771184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout, 139728976733056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['default_2'], 139722354727168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj, 139728976730704) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A, 139728976740064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'], 139722354724864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_A['default_2'].weight, 139726883359296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B, 139728976740112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['default_2'], 139722354725152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.base_layer, 139777663771136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout, 139728976731040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['default_2'], 139722354725104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out, 139728975375248) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A, 139728975374192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'], 139722354723808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_A['default_2'].weight, 139726883367376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B, 139728975376016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B['default_2'], 139722354723184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.base_layer, 139777663771376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout, 139728975374048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['default_2'], 139722354723328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k, 139777663771520) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_k.weight, 139777660416816) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q, 139777663771424) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.norm_added_q.weight, 139777664185616) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].attn.processor, 139777663770512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1, 139777663769840) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm, 139777663769984) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu, 139777663769888) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear, 139728976636384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A, 139728976638208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'], 139722354728128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_A['default_2'].weight, 139726883366336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B, 139728976638304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B['default_2'], 139722354731632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.base_layer, 139777663769936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout, 139728976636528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout['default_2'], 139722354729856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2, 139777663771568) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context, 139777663771904) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net, 139777663772048) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[8].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0], 139777663772000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj, 139728975414992) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A, 139728975413936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'], 139730397921824) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_A['default_2'].weight, 139726883367216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B, 139728975420080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['default_2'], 139730397932672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer, 139777663772096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout, 139728975429248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['default_2'], 139730397934016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1], 139777663772192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2], 139728976002496) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A, 139728976002592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'], 139730397926048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_A['default_2'].weight, 139726883359136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B, 139728976002928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B['default_2'], 139730398657552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].base_layer, 139777663772240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout, 139728976002112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['default_2'], 139730397927344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context, 139777663770560) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm, 139777663770032) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu, 139777663770224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear, 139728976635376) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[8].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A, 139728976639888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'], 139722354728704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_A['default_2'].weight, 139726883363776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B, 139728976640752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B['default_2'], 139722354728848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.base_layer, 139777663770368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout, 139728976627936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['default_2'], 139722354728272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[8].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[8].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[8].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[8].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context, 139777663771616) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[8].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9], 139777663769792) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff, 139777663773536) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net, 139777663773776) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0], 139777663773728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj, 139728975643552) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A, 139728975649696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'], 139722610931648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_A['default_2'].weight, 139727512433376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B, 139728975643696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['default_2'], 139722610933040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.base_layer, 139777663773824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout, 139728975645328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['default_2'], 139722612434704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1], 139777663773872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2], 139728975646576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A, 139728975647392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'], 139722607749312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_A['default_2'].weight, 139727512428496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B, 139728975646000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B['default_2'], 139722607754016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].base_layer, 139777663773920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout, 139728975646096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout['default_2'], 139722608938912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn, 139777663772768) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k, 139728977399040) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A, 139728977397984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'], 139724989829216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_A['default_2'].weight, 139727350099504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B, 139728977397264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B['default_2'], 139724974283680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.base_layer, 139777663772912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout, 139728977396064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout['default_2'], 139724979692032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q, 139728977410176) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A, 139728977405760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'], 139724983790144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_A['default_2'].weight, 139727354126448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B, 139728977407248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B['default_2'], 139724983788032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.base_layer, 139777663773008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout, 139728977409408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout['default_2'], 139724985784528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v, 139728977397312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A, 139728977399904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'], 139724976280944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_A['default_2'].weight, 139727350089744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B, 139728977397552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B['default_2'], 139724976290400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.base_layer, 139777663773104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout, 139728977409648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout['default_2'], 139724974292368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k, 139777663772960) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_k.weight, 139777664576272) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q, 139777663772864) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_q.weight, 139777664576352) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out, 139777663773296) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0], 139728975767792) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A, 139728975606592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'], 139722656150480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_A['default_2'].weight, 139732786054464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B, 139728975600592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B['default_2'], 139722656142512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].base_layer, 139777663773344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout, 139728975597616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['default_2'], 139722528145024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1], 139777663773392) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj, 139728975547024) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A, 139728975556864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'], 139728944545072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_A['default_2'].weight, 139722336804816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B, 139728975549760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['default_2'], 139728944533120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.base_layer, 139777663773152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout, 139728975549280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['default_2'], 139726561678320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj, 139728975769184) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A, 139728975771920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'], 139722528142336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_A['default_2'].weight, 139722336806576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B, 139728975768608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['default_2'], 139722528144400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.base_layer, 139777663773248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout, 139728975757760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['default_2'], 139722528144688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj, 139728975769328) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A, 139728975759536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'], 139722528138304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_A['default_2'].weight, 139722336801536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B, 139728975767312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['default_2'], 139722528136096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.base_layer, 139777663773200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout, 139728975768464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['default_2'], 139728944543056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out, 139728975602704) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A, 139728975647584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'], 139722612435424) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_A['default_2'].weight, 139727512435216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B, 139728975647488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B['default_2'], 139722612441040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.base_layer, 139777663773440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout, 139728975647728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['default_2'], 139722656147696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k, 139777663773584) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_k.weight, 139777664177936) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q, 139777663773488) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.norm_added_q.weight, 139777664576992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].attn.processor, 139777663772720) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1, 139777663772288) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm, 139777663772432) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu, 139777663772336) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear, 139728977408160) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A, 139728977408112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'], 139724972181152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_A['default_2'].weight, 139726883366096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B, 139728977407968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B['default_2'], 139724972180096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.base_layer, 139777663772384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout, 139728977408208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout['default_2'], 139730398656640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2, 139777663773632) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context, 139777663773968) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net, 139777663774112) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[9].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0], 139777663774064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj, 139728975646960) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A, 139728975924768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'], 139733025753744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_A['default_2'].weight, 139727512421296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B, 139728975936864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['default_2'], 139729096456128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer, 139777663774160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout, 139728975647680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['default_2'], 139722607753824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1], 139777663774256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2], 139728975897952) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A, 139728975900880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'], 139733030657072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_A['default_2'].weight, 139727512425056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B, 139728975899680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B['default_2'], 139733030659040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].base_layer, 139777663774304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout, 139728975898720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['default_2'], 139729096234736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context, 139777663772480) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm, 139777663772672) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu, 139777663772576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear, 139728977403168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[9].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A, 139728977405184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'], 139724970144000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_A['default_2'].weight, 139732786313344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B, 139728977407392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B['default_2'], 139724985785440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.base_layer, 139777663772624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout, 139728977405232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['default_2'], 139724972180192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[9].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[9].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[9].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[9].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context, 139777663773680) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[9].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10], 139777663771952) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff, 139777663775600) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net, 139777663775840) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0], 139777663775792) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj, 139728976051504) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A, 139728976046176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'], 139722743883680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_A['default_2'].weight, 139732787486112) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B, 139728976049488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['default_2'], 139722743880224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.base_layer, 139777663775888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout, 139728976037632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['default_2'], 139722743874896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1], 139777663775936) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2], 139728976049872) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A, 139728974655024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'], 139724951248976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_A['default_2'].weight, 139727510503264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B, 139728974655600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B['default_2'], 139722764636304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].base_layer, 139777663775984) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout, 139728974658480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout['default_2'], 139722767562880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn, 139777663774832) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k, 139728976191376) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A, 139728976190992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'], 139762699903680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_A['default_2'].weight, 139733010755280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B, 139728976192624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B['default_2'], 139762699904304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.base_layer, 139777663774976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout, 139728976191424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout['default_2'], 139762699902432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q, 139728975895216) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A, 139728975898432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'], 139762699904496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_A['default_2'].weight, 139728981460768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B, 139728976191280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B['default_2'], 139762699906032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.base_layer, 139777663775072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout, 139728975902272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout['default_2'], 139762699903392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v, 139728976199008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A, 139728976192576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'], 139762699902336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_A['default_2'].weight, 139733010744960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B, 139728976190896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B['default_2'], 139733030977936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.base_layer, 139777663775168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout, 139728976199152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout['default_2'], 139762699904688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k, 139777663775024) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_k.weight, 139777664181216) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q, 139777663774928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_q.weight, 139777664573792) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out, 139777663775360) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0], 139728977387696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A, 139728977382080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'], 139722746850288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_A['default_2'].weight, 139733010751200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B, 139728977379872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B['default_2'], 139722746852640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].base_layer, 139777663775408) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout, 139728977380976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['default_2'], 139722746852736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1], 139777663775456) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj, 139728976188112) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A, 139728976199536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'], 139722621712080) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_A['default_2'].weight, 139733010752960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B, 139728976197376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['default_2'], 139722621713904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.base_layer, 139777663775216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout, 139728976184032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['default_2'], 139733030981632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj, 139728975372992) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A, 139728976074912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'], 139722651029552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_A['default_2'].weight, 139733010748560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B, 139728976072368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['default_2'], 139722651025520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.base_layer, 139777663775312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout, 139728976081152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['default_2'], 139722651031568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj, 139728976188592) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A, 139728976184896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'], 139722621712896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_A['default_2'].weight, 139733010751440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B, 139728976190800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['default_2'], 139722655777408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.base_layer, 139777663775264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout, 139728976190416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['default_2'], 139722621712752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out, 139728975945472) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A, 139728975951952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'], 139722743883104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_A['default_2'].weight, 139733010758560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B, 139728975949888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B['default_2'], 139722743874512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.base_layer, 139777663775504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout, 139728975946672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['default_2'], 139722746839536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k, 139777663775648) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_k.weight, 139777664181776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q, 139777663775552) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.norm_added_q.weight, 139777664181136) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].attn.processor, 139777663774784) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1, 139777663774352) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm, 139777663774496) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu, 139777663774400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear, 139728975895840) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A, 139728975896704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'], 139733030659328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_A['default_2'].weight, 139727512420736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B, 139728975895888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B['default_2'], 139733030658704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.base_layer, 139777663774448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout, 139728975895600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout['default_2'], 139733030654096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2, 139777663775696) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context, 139777663776032) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net, 139777663776176) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[10].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0], 139777663776128) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj, 139728974653584) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A, 139728974644032) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'], 139722767982000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_A['default_2'].weight, 139727510493664) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B, 139728974653632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['default_2'], 139722767989056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer, 139777663776224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout, 139728974651280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['default_2'], 139722764636688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1], 139777663776320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2], 139728974651664) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A, 139728974657760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'], 139722528094960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_A['default_2'].weight, 139727510495024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B, 139728974647872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B['default_2'], 139722528094720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].base_layer, 139777663776368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout, 139728974645808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['default_2'], 139722528084880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context, 139777663774544) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm, 139777663774736) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu, 139777663774640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear, 139728975896224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[10].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A, 139728975898672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'], 139776725844032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_A['default_2'].weight, 139728981459488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B, 139728975898240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B['default_2'], 139762700247504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.base_layer, 139777663774688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout, 139728975896848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['default_2'], 139733030661872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[10].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[10].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[10].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[10].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context, 139777663775744) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[10].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11], 139777663774016) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff, 139777663777664) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net, 139777663777904) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0], 139777663777856) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj, 139728974535152) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A, 139728974534960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'], 139726567358032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_A['default_2'].weight, 139727510851632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B, 139728974530784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['default_2'], 139726567357216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.base_layer, 139777663777952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout, 139728974536208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['default_2'], 139724968807632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1], 139777663778000) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2], 139728974533472) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A, 139728974536640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'], 139724973570272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_A['default_2'].weight, 139727510848352) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B, 139728974538512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B['default_2'], 139722606841184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].base_layer, 139777663778048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout, 139728974538080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout['default_2'], 139724973557696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn, 139777663776896) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k, 139728975482784) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A, 139728975482112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'], 139722762551408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_A['default_2'].weight, 139727511132400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B, 139728975483456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B['default_2'], 139722762558896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.base_layer, 139777663777040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout, 139728975482064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout['default_2'], 139722762552896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q, 139728977032192) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A, 139728977025616) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'], 139722762562736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_A['default_2'].weight, 139727511140560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B, 139728977026192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B['default_2'], 139722762566384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.base_layer, 139777663777136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout, 139728977032480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout['default_2'], 139722525701504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v, 139728975486240) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A, 139728975481776) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'], 139722607493376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_A['default_2'].weight, 139727511130400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B, 139728975484464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B['default_2'], 139722607494096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.base_layer, 139777663777232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout, 139728975485520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout['default_2'], 139722609185248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k, 139777663777088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_k.weight, 139777664182176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q, 139777663776992) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_q.weight, 139777664571232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out, 139777663777424) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0], 139728976310288) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A, 139728976308608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'], 139722612136528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_A['default_2'].weight, 139726886403936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B, 139728976310480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B['default_2'], 139722612141616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].base_layer, 139777663777472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout, 139728976301216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['default_2'], 139722612136096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1], 139777663777520) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj, 139728975480336) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A, 139728975495072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'], 139722612137872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_A['default_2'].weight, 139727511132160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B, 139728975494016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['default_2'], 139722612139696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.base_layer, 139777663777280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout, 139728975480528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['default_2'], 139722612138160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj, 139728977346128) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A, 139728977337728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'], 139722612132688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_A['default_2'].weight, 139727511139600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B, 139728977332448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['default_2'], 139722612131776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.base_layer, 139777663777376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout, 139728977344592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['default_2'], 139722612134752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj, 139728976133760) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A, 139728976122096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'], 139722612141040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_A['default_2'].weight, 139727511139280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B, 139728976122336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['default_2'], 139722612140800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.base_layer, 139777663777328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout, 139728976133616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['default_2'], 139722612136480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out, 139728974536352) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A, 139728974538176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'], 139722612144016) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_A['default_2'].weight, 139727510858192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B, 139728974537120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B['default_2'], 139722612135136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.base_layer, 139777663777568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout, 139728974537264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['default_2'], 139722612143728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k, 139777663777712) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_k.weight, 139777664571952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q, 139777663777616) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.norm_added_q.weight, 139777664178656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].attn.processor, 139777663776848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1, 139777663776416) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm, 139777663776560) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu, 139777663776464) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear, 139728974655648) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A, 139728977035216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'], 139722528092032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_A['default_2'].weight, 139727510499584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B, 139728977021728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B['default_2'], 139722528081232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.base_layer, 139777663776512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout, 139728977022736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout['default_2'], 139722528093760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2, 139777663777760) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context, 139777663778096) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net, 139777663778240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[11].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0], 139777663778192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj, 139728974494400) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A, 139728974492960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'], 139722610579328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_A['default_2'].weight, 139727510857472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B, 139728974484848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['default_2'], 139722610579376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer, 139777663778288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout, 139728974490608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['default_2'], 139722606847808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1], 139777663778384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2], 139728976330320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A, 139728976322448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'], 139724981037152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_A['default_2'].weight, 139727510859872) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B, 139728976318800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B['default_2'], 139724981035520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].base_layer, 139777663778432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout, 139728976322016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['default_2'], 139730401330224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context, 139777663776608) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm, 139777663776800) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu, 139777663776704) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear, 139728977020480) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[11].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A, 139728977028160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'], 139722528094144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_A['default_2'].weight, 139727510502704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B, 139728977034304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B['default_2'], 139722528086080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.base_layer, 139777663776752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout, 139728977020816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['default_2'], 139722528091552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[11].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[11].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[11].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[11].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context, 139777663777808) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[11].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12], 139777663776080) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff, 139777663615952) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net, 139777663616240) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0], 139777663616192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj, 139728974844480) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A, 139728974849856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'], 139722358561168) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_A['default_2'].weight, 139727511862096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B, 139728974850864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['default_2'], 139722358561792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.base_layer, 139777663616288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout, 139728974847936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['default_2'], 139722358561360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1], 139777663616336) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2], 139728974741360) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A, 139728974736608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'], 139722358559536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_A['default_2'].weight, 139732785415168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B, 139728974739248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B['default_2'], 139722358560496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].base_layer, 139777663616384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout, 139728974738432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout['default_2'], 139722358560592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn, 139777663615184) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k, 139728974553984) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A, 139728974556096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'], 139722358556464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_A['default_2'].weight, 139724966146464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B, 139728973206176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B['default_2'], 139722358556992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.base_layer, 139777663615328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout, 139728974559984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout['default_2'], 139722358556416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q, 139728974560128) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A, 139728974546784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'], 139722358556176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_A['default_2'].weight, 139724966136624) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B, 139728974552976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B['default_2'], 139722358556272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.base_layer, 139777663615424) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout, 139728974551872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout['default_2'], 139722358556224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v, 139728973205312) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A, 139728973207328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'], 139722358565056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_A['default_2'].weight, 139724966144144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B, 139728973205696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B['default_2'], 139722358569376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.base_layer, 139777663615520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout, 139728973214144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout['default_2'], 139722358566448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k, 139777663615376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_k.weight, 139777664568592) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q, 139777663615280) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_q.weight, 139777664569312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out, 139777663615712) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0], 139728974829920) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A, 139728974830064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'], 139722358562176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_A['default_2'].weight, 139727511851936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B, 139728974836784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B['default_2'], 139722358562704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].base_layer, 139777663615760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout, 139728974823536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['default_2'], 139722358562800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1], 139777663615808) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj, 139728973211984) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A, 139728973209824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'], 139722358564576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_A['default_2'].weight, 139727511857776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B, 139728973202960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['default_2'], 139722358564624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.base_layer, 139777663615568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout, 139728973216688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['default_2'], 139722358564048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj, 139728976184800) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A, 139728974830736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'], 139722358562896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_A['default_2'].weight, 139727511847696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B, 139728974832464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['default_2'], 139722358563184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.base_layer, 139777663615664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout, 139728974827232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['default_2'], 139722358563136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj, 139728973215776) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A, 139728973207664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'], 139722358563808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_A['default_2'].weight, 139727511850896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B, 139728973207520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['default_2'], 139722358563232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.base_layer, 139777663615616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout, 139728973202240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['default_2'], 139722358563904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out, 139728975500656) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A, 139728974855712) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'], 139722358561888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_A['default_2'].weight, 139727511853216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B, 139728974847168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B['default_2'], 139722358561984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.base_layer, 139777663615856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout, 139728975507472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['default_2'], 139722358561840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k, 139777663616000) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_k.weight, 139777660406176) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q, 139777663615904) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.norm_added_q.weight, 139777664572832) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].attn.processor, 139777663615136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1, 139777663778480) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm, 139777663778624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu, 139777663778528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear, 139728974328640) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A, 139728974331376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'], 139722358555744) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_A['default_2'].weight, 139730405954304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B, 139728974322304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B['default_2'], 139722358556752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.base_layer, 139777663778576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout, 139728974325616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout['default_2'], 139724981733936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2, 139777663616048) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context, 139777663616432) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net, 139777663616576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[12].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0], 139777663616528) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj, 139728974740688) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A, 139728974741264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'], 139722358559488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_A['default_2'].weight, 139727512351376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B, 139728974737568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['default_2'], 139722358559728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer, 139777663616624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout, 139728974735936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['default_2'], 139722358559680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1], 139777663616720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2], 139728974729456) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A, 139728974726960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'], 139722607184480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_A['default_2'].weight, 139727512352656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B, 139728974728784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B['default_2'], 139722610048496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].base_layer, 139777663616768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout, 139728974726384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['default_2'], 139722608590320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context, 139777663778672) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm, 139777663615088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu, 139777663778768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear, 139728974461008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[12].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A, 139728974456928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'], 139722358555456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_A['default_2'].weight, 139730405959344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B, 139728974451696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B['default_2'], 139722358555696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.base_layer, 139777663615040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout, 139728974460768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['default_2'], 139722358555648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[12].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[12].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[12].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[12].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context, 139777663616096) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[12].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13], 139777663778144) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff, 139777663618064) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net, 139777663618304) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0], 139777663618256) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj, 139728973424832) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A, 139728973418592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'], 139733034545360) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_A['default_2'].weight, 139730407084880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B, 139728973425600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['default_2'], 139733034533264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.base_layer, 139777663618352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout, 139728973426080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['default_2'], 139733025049088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1], 139777663618400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2], 139728973423056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A, 139728973419792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'], 139728961030080) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_A['default_2'].weight, 139727511266592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B, 139728973423200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B['default_2'], 139728961037760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].base_layer, 139777663618448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout, 139728973418352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout['default_2'], 139733010959520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn, 139777663617296) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k, 139728974585456) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A, 139728974587664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'], 139724974250576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_A['default_2'].weight, 139727512352176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B, 139728974591264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B['default_2'], 139724974250624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.base_layer, 139777663617440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout, 139728974592416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout['default_2'], 139724969449776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q, 139728974729072) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A, 139728974726192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'], 139724970790768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_A['default_2'].weight, 139727512346576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B, 139728974725280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B['default_2'], 139724970795328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.base_layer, 139777663617536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout, 139728974728400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout['default_2'], 139728963801904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v, 139728974579072) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A, 139728974582048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'], 139728960000048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_A['default_2'].weight, 139727512352256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B, 139728974591552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B['default_2'], 139728960000960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.base_layer, 139777663617632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout, 139728974592704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout['default_2'], 139730399357072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k, 139777663617488) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_k.weight, 139777659967648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q, 139777663617392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_q.weight, 139777659967728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out, 139777663617824) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0], 139728974382944) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A, 139728974390192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'], 139722611269952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_A['default_2'].weight, 139727510669888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B, 139728974391440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B['default_2'], 139722611264384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].base_layer, 139777663617872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout, 139728974381696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['default_2'], 139724985098896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1], 139777663617920) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj, 139728974593376) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A, 139728974578880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'], 139727602989568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_A['default_2'].weight, 139727510679648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B, 139728974587520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['default_2'], 139727602989760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.base_layer, 139777663617680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout, 139728974587280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['default_2'], 139722607772192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj, 139728974387312) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A, 139728974388752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'], 139728959932112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_A['default_2'].weight, 139727510671248) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B, 139728974392160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['default_2'], 139728959935280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.base_layer, 139777663617776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout, 139728974394272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['default_2'], 139724987857408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj, 139728974384672) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A, 139728974382896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'], 139733021422784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_A['default_2'].weight, 139727510673168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B, 139728974381984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['default_2'], 139724987857552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.base_layer, 139777663617728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout, 139728974383376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['default_2'], 139727602990624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out, 139728974394800) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A, 139728974393072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'], 139722768386752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_A['default_2'].weight, 139727510679408) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B, 139728974392400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B['default_2'], 139733025051392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.base_layer, 139777663617968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout, 139728974394368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['default_2'], 139724966741904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k, 139777663618112) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_k.weight, 139777664697824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q, 139777663618016) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.norm_added_q.weight, 139777659967488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].attn.processor, 139777663617248) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1, 139777663616816) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm, 139777663616960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu, 139777663616864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear, 139728974741024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A, 139728974738000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'], 139722608435712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_A['default_2'].weight, 139727512347376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B, 139728974729504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B['default_2'], 139722608430048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.base_layer, 139777663616912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout, 139728974741168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout['default_2'], 139722609901920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2, 139777663618160) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context, 139777663618496) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net, 139777663618640) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[13].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0], 139777663618592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj, 139728973426896) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A, 139728973459904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'], 139724978338928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_A['default_2'].weight, 139727511262752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B, 139728973449824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['default_2'], 139722649304272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer, 139777663618688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout, 139728973428240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['default_2'], 139722606494144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1], 139777663618784) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2], 139728973462400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A, 139728973449008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'], 139730398560512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_A['default_2'].weight, 139727511260672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B, 139728973451744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B['default_2'], 139722608333136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].base_layer, 139777663618832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout, 139728973451936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['default_2'], 139724983052816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context, 139777663617008) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm, 139777663617200) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu, 139777663617104) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear, 139728974738816) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[13].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A, 139728974739152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'], 139722666759104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_A['default_2'].weight, 139727512342496) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B, 139728974738624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B['default_2'], 139722666760352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.base_layer, 139777663617152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout, 139728974739392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['default_2'], 139724972875952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[13].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[13].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[13].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[13].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context, 139777663618208) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[13].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14], 139777663616480) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff, 139777663620128) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net, 139777663620368) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0], 139777663620320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj, 139728975127808) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A, 139728975122960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'], 139722356554480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_A['default_2'].weight, 139733004053744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B, 139728975122672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['default_2'], 139722356555584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.base_layer, 139777663620416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout, 139728975128480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['default_2'], 139722356543968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1], 139777663620464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2], 139728975120320) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A, 139728975119072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'], 139722356019296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_A['default_2'].weight, 139722378068752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B, 139728975119648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B['default_2'], 139722356021744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].base_layer, 139777663620512) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout, 139728975120416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout['default_2'], 139722355887984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn, 139777663619360) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k, 139728975048192) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A, 139728975050592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'], 139722354597872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_A['default_2'].weight, 139722263637552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B, 139728975043248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B['default_2'], 139722354593168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.base_layer, 139777663619504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout, 139728975040080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout['default_2'], 139722354592784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q, 139728975049152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A, 139728975048720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'], 139722354604832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_A['default_2'].weight, 139722263640752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B, 139728975049008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B['default_2'], 139722354591920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.base_layer, 139777663619600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout, 139728975049200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout['default_2'], 139722354605408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v, 139728975044208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A, 139728975041568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'], 139722354594800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_A['default_2'].weight, 139722263628752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B, 139728975045648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B['default_2'], 139722354602000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.base_layer, 139777663619696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout, 139728975046464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout['default_2'], 139722354601088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k, 139777663619552) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_k.weight, 139777659969328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q, 139777663619456) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_q.weight, 139777659969408) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out, 139777663619888) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0], 139728974689280) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A, 139728974684864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'], 139722356542864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_A['default_2'].weight, 139730408239920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B, 139728974692304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B['default_2'], 139722356550832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].base_layer, 139777663619936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout, 139728974688176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['default_2'], 139722356549824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1], 139777663619984) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj, 139728975046944) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A, 139728975044736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'], 139722354590864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_A['default_2'].weight, 139722263635552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B, 139728973391920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['default_2'], 139722354602960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.base_layer, 139777663619744) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout, 139728975050352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['default_2'], 139722354597056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj, 139728974688416) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A, 139728974686784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'], 139722356550736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_A['default_2'].weight, 139730408241360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B, 139728974681456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['default_2'], 139722356545696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.base_layer, 139777663619840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout, 139728974686208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['default_2'], 139722356346160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj, 139728974691680) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A, 139728974687024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'], 139722354595328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_A['default_2'].weight, 139722263632192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B, 139728974692208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['default_2'], 139722354591872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.base_layer, 139777663619792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout, 139728974690864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['default_2'], 139722354604880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out, 139728974682032) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A, 139728975127520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'], 139722356554096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_A['default_2'].weight, 139732784615952) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B, 139728975129584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B['default_2'], 139722356544016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.base_layer, 139777663620032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout, 139728974689712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['default_2'], 139722356555056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k, 139777663620176) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_k.weight, 139777659969168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q, 139777663620080) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.norm_added_q.weight, 139777659969248) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].attn.processor, 139777663619312) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1, 139777663618880) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm, 139777663619024) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu, 139777663618928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear, 139728973451408) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A, 139728973450208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'], 139722609783856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_A['default_2'].weight, 139727511260032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B, 139728973456064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B['default_2'], 139722609778192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.base_layer, 139777663618976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout, 139728973451024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout['default_2'], 139722608334960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2, 139777663620224) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context, 139777663620560) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net, 139777663620704) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[14].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0], 139777663620656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj, 139728973207376) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A, 139728976288144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'], 139722356021408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_A['default_2'].weight, 139727511294960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B, 139728976288096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['default_2'], 139722356020832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer, 139777663620752) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout, 139728976282912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['default_2'], 139722356022656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1], 139777663620848) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2], 139728976282816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A, 139728976282768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'], 139722359114144) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_A['default_2'].weight, 139727511293280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B, 139728976290016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B['default_2'], 139722359118848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].base_layer, 139777663620896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout, 139728976281808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['default_2'], 139722359114816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context, 139777663619072) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm, 139777663619264) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu, 139777663619168) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear, 139728973459232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[14].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A, 139728973452560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'], 139722354592976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_A['default_2'].weight, 139727511268592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B, 139728973450736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B['default_2'], 139722354592640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.base_layer, 139777663619216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout, 139728973451072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['default_2'], 139729105086608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[14].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[14].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[14].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[14].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context, 139777663620272) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[14].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15], 139777663618544) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff, 139777663622192) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net, 139777663622480) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0], 139777663622384) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj, 139728974814496) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A, 139728974813296) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'], 139726866170640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_A['default_2'].weight, 139724970166288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B, 139728973566528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['default_2'], 139726866170352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.base_layer, 139777663622528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout, 139728974812768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['default_2'], 139726866175584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1], 139777663622576) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2], 139728973577136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A, 139728973566384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'], 139726866170496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_A['default_2'].weight, 139724970165008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B, 139728973563360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B['default_2'], 139726866172080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].base_layer, 139777663622624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout, 139728973578000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout['default_2'], 139726866173520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn, 139777663621424) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k, 139728973901744) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A, 139728973905200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'], 139722357295952) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_A['default_2'].weight, 139727512190096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B, 139728973892144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B['default_2'], 139722357297104) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.base_layer, 139777663621568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout, 139728973904864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout['default_2'], 139722357296096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q, 139728973890080) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A, 139728973891376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'], 139722356906768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_A['default_2'].weight, 139727511305680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B, 139728973905728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B['default_2'], 139722357299792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.base_layer, 139777663621664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout, 139728973904192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout['default_2'], 139722356763056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v, 139728974970256) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A, 139728974970832) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'], 139722357295520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_A['default_2'].weight, 139727512186256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B, 139728974964256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B['default_2'], 139722357295808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.base_layer, 139777663621760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout, 139728974964304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout['default_2'], 139722357293696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k, 139777663621616) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_k.weight, 139777664583696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q, 139777663621520) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_q.weight, 139777664581776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out, 139777663621952) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0], 139728973325616) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A, 139728973322592) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'], 139726866170928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_A['default_2'].weight, 139727512185616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B, 139728973329792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B['default_2'], 139726866175824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].base_layer, 139777663622000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout, 139728973325712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['default_2'], 139726866175008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1], 139777663622048) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj, 139728974273008) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A, 139728974278576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'], 139722357041248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_A['default_2'].weight, 139727512180016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B, 139728974279008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['default_2'], 139722357042976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.base_layer, 139777663621808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout, 139728974279536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['default_2'], 139722357295760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj, 139728973329936) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A, 139728973328928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'], 139726866172560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_A['default_2'].weight, 139727512175456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B, 139728973331136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['default_2'], 139726866176544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.base_layer, 139777663621904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout, 139728973329120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['default_2'], 139722357043840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj, 139728975132944) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A, 139728973318560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'], 139722357040864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_A['default_2'].weight, 139727512182656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B, 139728973319472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['default_2'], 139722357044464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.base_layer, 139777663621856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout, 139728975180704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['default_2'], 139722357038944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out, 139728973316352) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A, 139728973317120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'], 139726866173184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_A['default_2'].weight, 139724970179888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B, 139728973316496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B['default_2'], 139726866171264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.base_layer, 139777663622096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout, 139728973324752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['default_2'], 139726866173232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k, 139777663622240) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_k.weight, 139777664574432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q, 139777663622144) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.norm_added_q.weight, 139777664746656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].attn.processor, 139777663621376) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1, 139777663620944) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm, 139777663621088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu, 139777663620992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear, 139728973306544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A, 139728973309520) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'], 139722358708576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_A['default_2'].weight, 139727511297280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B, 139728973309088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B['default_2'], 139722358704160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.base_layer, 139777663621040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout, 139728973299968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout['default_2'], 139722359380032) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2, 139777663622288) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context, 139777663622672) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net, 139777663622816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[15].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0], 139777663622768) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj, 139728973570896) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A, 139728973562208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'], 139726866170064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_A['default_2'].weight, 139724970169808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B, 139728973570128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['default_2'], 139726866175776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer, 139777663622864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout, 139728973570944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['default_2'], 139726866175536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1], 139777663622960) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2], 139728973565616) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A, 139728973576128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'], 139726866173712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_A['default_2'].weight, 139733014610384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B, 139728973564272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B['default_2'], 139726866173328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].base_layer, 139777663623008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout, 139728973565808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['default_2'], 139726866173376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context, 139777663621136) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm, 139777663621328) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu, 139777663621232) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear, 139728973890800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[15].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A, 139728973905488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'], 139722358711840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_A['default_2'].weight, 139727511303520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B, 139728973890128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B['default_2'], 139722356756960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.base_layer, 139777663621280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout, 139728973905344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['default_2'], 139722358711936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[15].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[15].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[15].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[15].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context, 139777663622336) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[15].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16], 139777663620608) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff, 139777663624352) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net, 139777663624592) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0], 139777663624544) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj, 139728972213568) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A, 139728974061024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'], 139726866177840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_A['default_2'].weight, 139733015839824) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B, 139728974066448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['default_2'], 139726866177696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.base_layer, 139777663624640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout, 139728974068176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['default_2'], 139726866177792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1], 139777663624688) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2], 139728974278192) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A, 139728973822512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'], 139726866178272) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_A['default_2'].weight, 139732790225680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B, 139728973823472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B['default_2'], 139726866178128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].base_layer, 139777663624736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout, 139728973810224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout['default_2'], 139726866178224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn, 139777663623584) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k, 139728973512464) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A, 139728973502192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'], 139726866173088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_A['default_2'].weight, 139727512459424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B, 139728973508960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B['default_2'], 139726866173952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.base_layer, 139777663623728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout, 139728973512224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout['default_2'], 139726866175248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q, 139728973510448) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A, 139728973499216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'], 139726866171648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_A['default_2'].weight, 139727512463424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B, 139728973500080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B['default_2'], 139726866176112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.base_layer, 139777663623824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout, 139728973500656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout['default_2'], 139726866174864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v, 139728972211408) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A, 139728972217504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'], 139726866171792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_A['default_2'].weight, 139727512458224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B, 139728972203872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B['default_2'], 139726866172800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.base_layer, 139777663623920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout, 139728972212512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout['default_2'], 139726866172944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k, 139777663623776) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_k.weight, 139777664327232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q, 139777663623680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_q.weight, 139777664594016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out, 139777663624112) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0], 139728972202096) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A, 139728972204352) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'], 139726866176976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_A['default_2'].weight, 139732790911488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B, 139728972203584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B['default_2'], 139726866176832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].base_layer, 139777663624160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout, 139728972204832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['default_2'], 139726866176928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1], 139777663624208) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj, 139728972209584) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A, 139728972204784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'], 139726866174384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_A['default_2'].weight, 139727512457984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B, 139728972205504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['default_2'], 139726866172176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.base_layer, 139777663623968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout, 139728972217264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['default_2'], 139726866174720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj, 139728972215824) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A, 139728972215200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'], 139726866176448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_A['default_2'].weight, 139732790914448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B, 139728972214912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['default_2'], 139726866170592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.base_layer, 139777663624064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout, 139728972216352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['default_2'], 139726866176688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj, 139728972216640) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A, 139728972218176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'], 139726866175200) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_A['default_2'].weight, 139727512456304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B, 139728972217648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['default_2'], 139726866172224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.base_layer, 139777663624016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout, 139728972209872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['default_2'], 139726866173616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out, 139728972211792) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A, 139728972213952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'], 139726866177408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_A['default_2'].weight, 139730405796544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B, 139728972212464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B['default_2'], 139726866177264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.base_layer, 139777663624256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout, 139728972216400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['default_2'], 139726866177360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k, 139777663624400) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_k.weight, 139777664592896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q, 139777663624304) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.norm_added_q.weight, 139777664595776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].attn.processor, 139777663623536) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1, 139777663623056) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm, 139777663623248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu, 139777663623152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear, 139728973566000) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A, 139728973569504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'], 139726866175152) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_A['default_2'].weight, 139733014613664) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B, 139728973563840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B['default_2'], 139726866174048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.base_layer, 139777663623200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout, 139728973569216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout['default_2'], 139726866171120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2, 139777663624448) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context, 139777663624784) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net, 139777663624928) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[16].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0], 139777663624880) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj, 139728975241872) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A, 139728975233280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'], 139726866178704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_A['default_2'].weight, 139730405220144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B, 139728975249216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['default_2'], 139726866178560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer, 139777663624976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout, 139728975246912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['default_2'], 139726866178656) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1], 139777663625072) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2], 139728972455344) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A, 139728972463312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'], 139726866179136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_A['default_2'].weight, 139732790527376) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B, 139728972452704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B['default_2'], 139726866178992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].base_layer, 139777663625120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout, 139728972451552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['default_2'], 139726866179088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context, 139777663623296) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm, 139777663623488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu, 139777663623392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear, 139728973578096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[16].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A, 139728973506272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'], 139726866175488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_A['default_2'].weight, 139727512467424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B, 139728973509152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B['default_2'], 139726866171216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.base_layer, 139777663623440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout, 139728973563456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['default_2'], 139726866171360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[16].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[16].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[16].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[16].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context, 139777663624496) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[16].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17], 139777663622720) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff, 139777663626416) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net, 139777663626656) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0], 139777663626608) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj, 139728974249120) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A, 139728974247968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'], 139726866183888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_A['default_2'].weight, 139722765728752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B, 139728974246144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['default_2'], 139726866183744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.base_layer, 139777663626704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout, 139728974248640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['default_2'], 139726866183840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1], 139777663626752) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2], 139728974246912) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A, 139728974246864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'], 139726866184320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_A['default_2'].weight, 139730407565056) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B, 139728974245952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B['default_2'], 139726866184176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].base_layer, 139777663626800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout, 139728974247008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout['default_2'], 139726866184272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn, 139777663625648) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k, 139728972369440) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A, 139728972366224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'], 139726866180864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_A['default_2'].weight, 139727512247712) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B, 139728972380480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B['default_2'], 139726866180720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.base_layer, 139777663625792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout, 139728972380288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout['default_2'], 139726866180816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q, 139728972371600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A, 139728972371024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'], 139726866180432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_A['default_2'].weight, 139727512254832) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B, 139728972375680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B['default_2'], 139726866180288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.base_layer, 139777663625888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout, 139728972378896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout['default_2'], 139726866180384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v, 139728973750064) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A, 139728973742720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'], 139726866181296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_A['default_2'].weight, 139727512244432) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B, 139728973750208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B['default_2'], 139726866181152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.base_layer, 139777663625984) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout, 139728973749632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout['default_2'], 139726866181248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k, 139777663625840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_k.weight, 139777675123168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q, 139777663625744) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_q.weight, 139777660095440) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out, 139777663626176) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0], 139728974236016) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A, 139728974237456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'], 139726866183024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_A['default_2'].weight, 139730405418592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B, 139728974236976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B['default_2'], 139726866182880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].base_layer, 139777663626224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout, 139728974236256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['default_2'], 139726866182976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1], 139777663626272) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj, 139728975195984) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A, 139728975194880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'], 139726866181728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_A['default_2'].weight, 139727512244752) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B, 139728975195696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['default_2'], 139726866181584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.base_layer, 139777663626032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout, 139728975192624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['default_2'], 139726866181680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj, 139728973716768) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A, 139728974235824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'], 139726866182592) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_A['default_2'].weight, 139730405419472) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B, 139728974236880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['default_2'], 139726866182448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.base_layer, 139777663626128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout, 139728974236496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['default_2'], 139726866182544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj, 139728973722000) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A, 139728973722240) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'], 139726866182160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_A['default_2'].weight, 139730403950976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B, 139728973714944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['default_2'], 139726866182016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.base_layer, 139777663626080) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout, 139728973721904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['default_2'], 139726866182112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out, 139728974237408) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A, 139728974247680) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'], 139726866183456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_A['default_2'].weight, 139733038599840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B, 139728974244848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B['default_2'], 139726866183312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.base_layer, 139777663626320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout, 139728974242208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['default_2'], 139726866183408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k, 139777663626464) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_k.weight, 139777664329392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q, 139777663626368) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.norm_added_q.weight, 139777664585456) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].attn.processor, 139777663625600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1, 139777663625168) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm, 139777663625312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu, 139777663625216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear, 139728972458800) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A, 139728972463744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'], 139726866179568) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_A['default_2'].weight, 139722618481904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B, 139728972374240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B['default_2'], 139726866179424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.base_layer, 139777663625264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout, 139728972453472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout['default_2'], 139726866179520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2, 139777663626512) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context, 139777663626848) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net, 139777663626992) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[17].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0], 139777663626944) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj, 139728974234960) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A, 139728974237216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'], 139726866184752) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_A['default_2'].weight, 139732786150944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B, 139728974235584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['default_2'], 139726866184608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer, 139777663627040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout, 139728974235680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['default_2'], 139726866184704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1], 139777663627136) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2], 139728974062464) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A, 139728975001632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'], 139726866185184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_A['default_2'].weight, 139733004856720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B, 139728974994192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B['default_2'], 139726866185040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].base_layer, 139777663627184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout, 139728975002064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['default_2'], 139726866185136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context, 139777663625360) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm, 139777663625552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu, 139777663625456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear, 139728972375728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[17].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A, 139728972376304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'], 139726866180000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_A['default_2'].weight, 139722618477744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B, 139728972372272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B['default_2'], 139726866179856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.base_layer, 139777663625504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout, 139728972371792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['default_2'], 139726866179952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[17].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[17].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[17].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[17].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context, 139777663626560) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[17].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18], 139777663624832) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff, accessed_by=DictGetItemGuardAccessor(ff) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff, 139777663628480) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.training, 139786895605728) # ff_output = self.ff(norm_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:185 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net, 139777663628720) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0], 139777663628672) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj, 139728972775184) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A, 139728972762416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'], 139727277684864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_A['default_2'].weight, 139727509268864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B, 139728972762320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['default_2'], 139727277575712) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.base_layer, 139777663628768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout, 139728972760784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['default_2'], 139727279404224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1], 139777663628816) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2], 139728974236064) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A, 139728972688512) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'], 139727275247600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_A['default_2'].weight, 139727509262304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B, 139728972691776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B['default_2'], 139727274008608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].base_layer, 139777663628864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout, 139728972682800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout['default_2'], 139727275959216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn, 139777663627712) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.training, 139786895605728) # attn_output, context_attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:172 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k, 139728973977088) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A, 139728973976896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'], 139727312443024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_A['default_2'].weight, 139727509259664) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B, 139728973975408) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B['default_2'], 139727312445520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.base_layer, 139777663627856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout, 139728973977136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout['default_2'], 139727312562464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q, 139728973973920) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A, 139728973979296) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'], 139727316447920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_A['default_2'].weight, 139727509266064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B, 139728973976464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B['default_2'], 139727320559056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.base_layer, 139777663627952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout, 139728973972816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout['default_2'], 139727314425296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v, 139728974183152) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A, 139728974168848) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'], 139727310289456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_A['default_2'].weight, 139727509273344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B, 139728974184256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B['default_2'], 139727310280048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.base_layer, 139777663628048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout, 139728974182144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout['default_2'], 139727312883376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k, 139777663627904) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_k.weight, 139777664592016) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q, 139777663627808) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_q.weight, 139777664592096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out, accessed_by=DictGetItemGuardAccessor(to_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out, 139777663628240) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out.training, 139786895605728) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0], 139728972766976) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_out[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].training, 139786895605696) # hidden_states = attn.to_out[0](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1776 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A, 139728972761744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'], 139727282290640) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_A['default_2'].weight, 139727509268224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B, 139728972767168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B['default_2'], 139727285566864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].base_layer, 139777663628288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout, 139728972774128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['default_2'], 139727283824016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_out[0].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_out[0].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_out[0].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_out[0].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[0]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_out[0]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1], 139777663628336) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_out[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_out[1].training, 139786895605728) # hidden_states = attn.to_out[1](hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1778 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj, accessed_by=DictGetItemGuardAccessor(add_k_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj, 139728974183296) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_k_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.training, 139786895605696) # encoder_hidden_states_key_proj = attn.add_k_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1736 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A, 139728974182672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'], 139727305401216) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_A['default_2'].weight, 139727509269104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B, 139728974169088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['default_2'], 139727305401840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.base_layer, 139777663628096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout, 139728974181616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['default_2'], 139727305551984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_k_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_k_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_k_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_k_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_k_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj, accessed_by=DictGetItemGuardAccessor(add_q_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj, 139728972766208) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_q_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.training, 139786895605696) # encoder_hidden_states_query_proj = attn.add_q_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1735 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A, 139728972767552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'], 139727292657968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_A['default_2'].weight, 139727509260384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B, 139728972767600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['default_2'], 139727293339856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.base_layer, 139777663628192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout, 139728972766592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['default_2'], 139727293747872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_q_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_q_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_q_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_q_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_q_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj, accessed_by=DictGetItemGuardAccessor(add_v_proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj, 139728974173072) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.add_v_proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.training, 139786895605696) # encoder_hidden_states_value_proj = attn.add_v_proj(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1737 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A, 139728974173504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'], 139727301111392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_A['default_2'].weight, 139727509270704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B, 139728972774080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['default_2'], 139727301105824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.base_layer, 139777663628144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout, 139728974178928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['default_2'], 139727304755616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_v_proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.add_v_proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.add_v_proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.add_v_proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.add_v_proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out, accessed_by=DictGetItemGuardAccessor(to_add_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out, 139728972767888) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.to_add_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.training, 139786895605696) # encoder_hidden_states = attn.to_add_out(encoder_hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1779 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A, 139728972760208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'], 139727279823680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_A['default_2'].weight, 139727509261264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B, 139728972768608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B['default_2'], 139727281801680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.base_layer, 139777663628384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout, 139728972760544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['default_2'], 139727281032016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_add_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.to_add_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].attn.to_add_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].attn.to_add_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.to_add_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.to_add_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].attn.to_add_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k, accessed_by=DictGetItemGuardAccessor(norm_added_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k, 139777663628528) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.training, 139786895605728) # if attn.norm_added_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1751 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_k.weight, 139777664758736) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q, accessed_by=DictGetItemGuardAccessor(norm_added_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q, 139777663628432) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].attn.norm_added_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.training, 139786895605728) # if attn.norm_added_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1749 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.norm_added_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.norm_added_q.weight, 139777664333392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.norm_added_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].attn.processor, 139777663627664) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1, accessed_by=DictGetItemGuardAccessor(norm1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1, 139777663627232) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.training, 139786895605728) # norm_hidden_states, gate_msa, shift_mlp, scale_mlp, gate_mlp = self.norm1(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:165 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm, 139777663627376) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu, 139777663627280) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear, 139728974995488) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A, 139728973972048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'], 139726866185616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_A['default_2'].weight, 139733043248672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B, 139728973971904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B['default_2'], 139726866185472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.base_layer, 139777663627328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout, 139728973978864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout['default_2'], 139726866185568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].norm1.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].norm1.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2, accessed_by=DictGetItemGuardAccessor(norm2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2, 139777663628576) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2.training, 139786895605728) # norm_hidden_states = self.norm2(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:182 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context, accessed_by=DictGetItemGuardAccessor(ff_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context, 139777663628912) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.training, 139786895605728) # context_ff_output = self.ff_context(norm_encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:198 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net, accessed_by=DictGetItemGuardAccessor(net) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net, 139777663629056) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net, 94649989240736) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- LENGTH_CHECK: len(L['self'].transformer_blocks[18].ff_context.net) == 3 # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net.training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0], 139777663629008) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj, accessed_by=DictGetItemGuardAccessor(proj) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj, 139728972759776) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[0].proj.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.training, 139786895605696) # hidden_states = self.proj(hidden_states) # diffusers/src/diffusers/models/activations.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A, 139728972501184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'], 139727272094656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_A['default_2'].weight, 139727509267904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B, 139728972510736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['default_2'], 139727272672512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer, 139777663629104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout, 139728972512896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['default_2'], 139727272515936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].proj.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff_context.net[0].proj.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[0].proj._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[0].proj._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0].approximate, accessed_by=DictGetItemGuardAccessor(approximate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[0].approximate == 'tanh' # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1], 139777663629200) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[1].training, 139786895605728) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2], 139728972498400) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].ff_context.net[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].training, 139786895605696) # for module in self.net: # diffusers/src/diffusers/models/attention.py:1200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A, 139728972505792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'], 139727260204528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_A['default_2'].weight, 139727509267024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B, 139728972500464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B['default_2'], 139727266211024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].base_layer, 139777663629248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout, 139728972510160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['default_2'], 139727261335616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[2].scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].ff_context.net[2].scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].ff_context.net[2].use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2].use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].ff_context.net[2].merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].ff_context.net[2]._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].ff_context.net[2]._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].ff_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context, accessed_by=DictGetItemGuardAccessor(norm1_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context, 139777663627424) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.training, 139786895605728) # norm_encoder_hidden_states, c_gate_msa, c_shift_mlp, c_scale_mlp, c_gate_mlp = self.norm1_context( # diffusers/src/diffusers/models/transformers/transformer_flux.py:167 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.emb, accessed_by=DictGetItemGuardAccessor(emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.emb, 139786895690912) # if self.emb is not None: # diffusers/src/diffusers/models/normalization.py:135 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm, 139777663627616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:139 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu, 139777663627520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear, 139728973986400) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].transformer_blocks[18].norm1_context.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:137 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A, 139728973986208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'], 139726866186048) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_A['default_2'].weight, 139727509263184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B, 139728973986160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B['default_2'], 139726866185904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.base_layer, 139777663627568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout, 139728973986688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['default_2'], 139726866186000) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1_context.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].transformer_blocks[18].norm1_context.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].transformer_blocks[18].norm1_context.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].transformer_blocks[18].norm1_context.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm1_context.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].transformer_blocks[18].norm1_context.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm1_context._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context, accessed_by=DictGetItemGuardAccessor(norm2_context) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context, 139777663628624) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18].norm2_context.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].transformer_blocks[18].norm2_context.training, 139786895605728) # norm_encoder_hidden_states = self.norm2_context(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:195 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=L['self'].single_transformer_blocks, accessed_by=DictGetItemGuardAccessor(single_transformer_blocks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks, 139777663628960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks.training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0], 139777663626896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn, 139777663629776) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k, 139728973631584) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A, 139728973630048) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'], 139724965827184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_A['default_2'].weight, 139727509269584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B, 139728973628368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B['default_2'], 139724965826416) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.base_layer, 139777663629920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout, 139728973631632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['default_2'], 139724965825552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q, 139728975203488) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A, 139728975210112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'], 139726888726880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_A['default_2'].weight, 139727509271984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B, 139728975214384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B['default_2'], 139726888726496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.base_layer, 139777663630016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout, 139728975203872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['default_2'], 139726890891536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v, 139728973627696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A, 139728973629568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'], 139724965824160) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_A['default_2'].weight, 139727509266944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B, 139728973629664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B['default_2'], 139724965826464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.base_layer, 139777663630064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout, 139728973628176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['default_2'], 139727316877664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k, 139777663629968) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_k.weight, 139777659813168) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q, 139777663629824) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.norm_q.weight, 139777660410496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.processor, 139777663629728) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm, 139777663629344) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm, 139777663629488) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu, 139777663629392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear, 139728972497728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A, 139728972498064) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'], 139727253074848) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_A['default_2'].weight, 139727509262704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B, 139728972498736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B['default_2'], 139727253071632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.base_layer, 139777663629440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout, 139728972497248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['default_2'], 139727258430832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp, 139777663629632) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp, 139728972501952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A, 139728972504400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'], 139726882410352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_A['default_2'].weight, 139727509269184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B, 139728972512464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B['default_2'], 139727236367504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.base_layer, 139777663629584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout, 139728972510640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['default_2'], 139727235572896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out, 139728975204064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[0].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A, 139728975202480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'], 139726893730016) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_A['default_2'].weight, 139727509273184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B, 139728975204304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B['default_2'], 139727248194432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.base_layer, 139777663629680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout, 139728975202432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout['default_2'], 139726882567408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[0].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[0].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[0].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[0].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[0].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[0].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[0]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1], 139777663629296) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn, 139777663630544) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k, 139728974145968) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A, 139728974141264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'], 139722333258560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_A['default_2'].weight, 139727509258304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B, 139728974093744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B['default_2'], 139722333269744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.base_layer, 139777663630688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout, 139728974147696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['default_2'], 139722333599296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q, 139728973400240) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A, 139728973410608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'], 139727246002304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_A['default_2'].weight, 139727509272304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B, 139728974137520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B['default_2'], 139727246014832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.base_layer, 139777663630784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout, 139728973400336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['default_2'], 139722333101200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v, 139728974100320) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A, 139728974092016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'], 139727236822208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_A['default_2'].weight, 139727509268544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B, 139728974102240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B['default_2'], 139727240520768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.base_layer, 139777663630832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout, 139728974100608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['default_2'], 139722333523392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k, 139777663630736) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_k.weight, 139777660121344) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q, 139777663630592) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.norm_q.weight, 139777664173856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.processor, 139777663630496) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm, 139777663630160) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm, 139777663630304) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu, 139777663630208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear, 139728973632112) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A, 139728973630864) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'], 139727315692912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_A['default_2'].weight, 139727509264144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B, 139728973632160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B['default_2'], 139727316993552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.base_layer, 139777663630256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout, 139728973643392) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['default_2'], 139727314813568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp, 139777663630400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp, 139728973863936) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A, 139728973866144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'], 139722764372672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_A['default_2'].weight, 139727509263744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B, 139728973863024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B['default_2'], 139727316667888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.base_layer, 139777663630352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout, 139728973859856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['default_2'], 139727312024336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out, 139728973410368) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[1].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A, 139728973401008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'], 139727242809536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_A['default_2'].weight, 139727509262224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B, 139728973399616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B['default_2'], 139727251589392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.base_layer, 139777663630448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout, 139728973406768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout['default_2'], 139727288318512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[1].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[1].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[1].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[1].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[1].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[1].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[1]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2], 139777663630112) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn, 139777663631312) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k, 139728972985248) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A, 139728972986688) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'], 139726888543920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_A['default_2'].weight, 139727509266304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B, 139728972986976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B['default_2'], 139726888545888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.base_layer, 139777662337184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout, 139728972979296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['default_2'], 139726888546080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q, 139728972288048) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A, 139728972295200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'], 139726888543728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_A['default_2'].weight, 139727509270464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B, 139728972290448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B['default_2'], 139726888538832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.base_layer, 139777662337280) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout, 139728972292320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['default_2'], 139726888544256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v, 139728972974880) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A, 139728972976896) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'], 139726888537296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_A['default_2'].weight, 139727509274384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B, 139728972984528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B['default_2'], 139726888549536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.base_layer, 139777662337328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout, 139728972986352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['default_2'], 139726888547712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k, 139777662337232) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_k.weight, 139777660414656) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q, 139777662337088) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.norm_q.weight, 139777664172096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.processor, 139777663631264) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm, 139777663630928) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm, 139777663631072) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu, 139777663630976) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear, 139728972338208) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A, 139728972730464) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'], 139726888077632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_A['default_2'].weight, 139727509262784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B, 139728972742080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B['default_2'], 139726888079744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.base_layer, 139777663631024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout, 139728972335136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['default_2'], 139726888646688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp, 139777663631168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp, 139728972727200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A, 139728972296112) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'], 139726888080992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_A['default_2'].weight, 139727509258784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B, 139728972287280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B['default_2'], 139726888076912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.base_layer, 139777663631120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout, 139728972286176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['default_2'], 139726888080704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out, 139728972288768) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[2].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A, 139728972287760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'], 139726888210816) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_A['default_2'].weight, 139727509270064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B, 139728972288816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B['default_2'], 139726888540848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.base_layer, 139777663631216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout, 139728972288384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout['default_2'], 139726888211200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[2].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[2].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[2].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[2].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[2].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[2].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[2]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3], accessed_by=GetItemGuardAccessor(3) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3], 139777663630880) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn, 139777662337808) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k, 139728973077600) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A, 139728973136992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'], 139726887932384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_A['default_2'].weight, 139727509270304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B, 139728973145872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B['default_2'], 139726887932480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.base_layer, 139777662337952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout, 139728973080816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['default_2'], 139726887932240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q, 139728973080432) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A, 139728973086624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'], 139726887932576) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_A['default_2'].weight, 139727509274544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B, 139728973082400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B['default_2'], 139726887933152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.base_layer, 139777662338048) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout, 139728973079136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['default_2'], 139726887932288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v, 139728973140496) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A, 139728973149616) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'], 139726887931808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_A['default_2'].weight, 139727509264624) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B, 139728973151872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B['default_2'], 139726887932096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.base_layer, 139777662338096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout, 139728973139872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['default_2'], 139726887931760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k, 139777662338000) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_k.weight, 139777660434480) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q, 139777662337856) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.norm_q.weight, 139777664183776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.processor, 139777662337760) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm, 139777662337424) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm, 139777662337568) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu, 139777662337472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear, 139728972975984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A, 139728972979200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'], 139726888537392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_A['default_2'].weight, 139727509270864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B, 139728972985968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B['default_2'], 139726888538928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.base_layer, 139777662337520) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout, 139728972983280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['default_2'], 139726888539840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp, 139777662337664) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp, 139728973081968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A, 139728973081488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'], 139726888545024) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_A['default_2'].weight, 139727509265024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B, 139728973082160) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B['default_2'], 139726887942080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.base_layer, 139777662337616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout, 139728973073184) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['default_2'], 139726888547856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out, 139728973078032) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[3].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A, 139728973077648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'], 139726887933920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_A['default_2'].weight, 139727509258944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B, 139728973078800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B['default_2'], 139726887933776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.base_layer, 139777662337712) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout, 139728973078992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout['default_2'], 139726887933680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[3].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[3].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[3].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[3].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[3].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[3].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[3]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4], accessed_by=GetItemGuardAccessor(4) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4], 139777662337376) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn, 139777662338576) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k, 139728972269024) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A, 139728972272432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'], 139726887928448) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_A['default_2'].weight, 139727353916416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B, 139728972271184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B['default_2'], 139726887928112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.base_layer, 139777662338720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout, 139728972277232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['default_2'], 139726887928160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q, 139728974045168) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A, 139728974038208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'], 139726887928832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_A['default_2'].weight, 139733026196192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B, 139728974042096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B['default_2'], 139726887929168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.base_layer, 139777662338816) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout, 139728974048432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['default_2'], 139726887928640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v, 139728972275936) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A, 139728972269072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'], 139726887835904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_A['default_2'].weight, 139727353912096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B, 139728972269648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B['default_2'], 139726887836096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.base_layer, 139777662338864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout, 139728972276752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['default_2'], 139726887927920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k, 139777662338768) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_k.weight, 139777664176976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q, 139777662338624) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.norm_q.weight, 139777660379648) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.processor, 139777662338528) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm, 139777662338192) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm, 139777662338336) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu, 139777662338240) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear, 139728973143520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A, 139728973140736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'], 139726887931328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_A['default_2'].weight, 139727351883360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B, 139728973144384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B['default_2'], 139726887930992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.base_layer, 139777662338288) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout, 139728973145680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['default_2'], 139726887931040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp, 139777662338432) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp, 139728973147696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A, 139728973147024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'], 139726887929696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_A['default_2'].weight, 139727358490992) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B, 139728973144144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B['default_2'], 139726887930752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.base_layer, 139777662338384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout, 139728973138240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['default_2'], 139726887930848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out, 139728973149664) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[4].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A, 139728973146016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'], 139726887929552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_A['default_2'].weight, 139727357704960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B, 139728973146832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B['default_2'], 139726887930080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.base_layer, 139777662338480) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout, 139728973152160) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout['default_2'], 139726887929360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[4].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[4].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[4].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[4].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[4].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[4].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[4]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5], accessed_by=GetItemGuardAccessor(5) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5], 139777662338144) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn, 139777662339344) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k, 139728972551920) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A, 139728972551152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'], 139726887842624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_A['default_2'].weight, 139727353914176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B, 139728972553648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B['default_2'], 139726887842240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.base_layer, 139777662339488) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout, 139728972552592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['default_2'], 139726887842576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q, 139728973662624) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A, 139728972557104) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'], 139726887843296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_A['default_2'].weight, 139727353910576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B, 139728972553936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B['default_2'], 139726887842480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.base_layer, 139777662339584) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout, 139728972556336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['default_2'], 139726887843248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v, 139728972554992) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A, 139728972816272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'], 139726887840992) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_A['default_2'].weight, 139727353911296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B, 139728972813872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B['default_2'], 139726887841376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.base_layer, 139777662339632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout, 139728972558256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['default_2'], 139726887841520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k, 139777662339536) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_k.weight, 139781198839008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q, 139777662339392) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.norm_q.weight, 139777664579472) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.processor, 139777662339296) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm, 139777662338960) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm, 139777662339104) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu, 139777662339008) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear, 139728972282128) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A, 139728972268880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'], 139726887845408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_A['default_2'].weight, 139727353913296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B, 139728972270752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B['default_2'], 139726887843968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.base_layer, 139777662339056) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout, 139728972277568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['default_2'], 139726887845360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp, 139777662339200) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp, 139728972271328) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A, 139728972269888) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'], 139726887843776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_A['default_2'].weight, 139727353909856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B, 139728972276704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B['default_2'], 139726887844688) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.base_layer, 139777662339152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout, 139728972272144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['default_2'], 139726887844784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out, 139728973673328) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[5].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A, 139728973661136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'], 139726887843872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_A['default_2'].weight, 139727353909136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B, 139728973674384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B['default_2'], 139726887843920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.base_layer, 139777662339248) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout, 139728973676064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout['default_2'], 139726887843200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[5].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[5].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[5].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[5].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[5].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[5].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[5]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6], accessed_by=GetItemGuardAccessor(6) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6], 139777662338912) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn, 139777662340112) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k, 139728971307072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A, 139728971302176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'], 139726887830528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_A['default_2'].weight, 139727353917296) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B, 139728971307168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B['default_2'], 139726887830384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.base_layer, 139777662340256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout, 139728971310960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['default_2'], 139726887831296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q, 139728971315280) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A, 139728971314416) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'], 139726887830480) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_A['default_2'].weight, 139727353906016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B, 139728971308080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B['default_2'], 139726887838640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.base_layer, 139777662340352) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout, 139728971312352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['default_2'], 139726887830288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v, 139728971204736) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A, 139728971213184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'], 139726887832112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_A['default_2'].weight, 139727353911456) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B, 139728971212224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B['default_2'], 139726887837008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.base_layer, 139777662340400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout, 139728971209968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['default_2'], 139726887830624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k, 139777662340304) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_k.weight, 139777664186496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q, 139777662340160) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.norm_q.weight, 139777660410336) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.processor, 139777662340064) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm, 139777662339728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm, 139777662339872) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu, 139777662339776) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear, 139728972812576) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A, 139728972811568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'], 139726887840800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_A['default_2'].weight, 139727353914256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B, 139728972811376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B['default_2'], 139726887840896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.base_layer, 139777662339824) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout, 139728972813008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['default_2'], 139726887840752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp, 139777662339968) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp, 139728972818288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A, 139728972818000) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'], 139726887840080) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_A['default_2'].weight, 139727353905536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B, 139728972623360) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B['default_2'], 139726887840704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.base_layer, 139777662339920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout, 139728972818192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['default_2'], 139726887840272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out, 139728972618224) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[6].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A, 139728971314128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'], 139726887839504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_A['default_2'].weight, 139727353914896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B, 139728971310624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B['default_2'], 139726887839120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.base_layer, 139777662340016) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout, 139728971302416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout['default_2'], 139726887839456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[6].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[6].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[6].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[6].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[6].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[6].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[6]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7], accessed_by=GetItemGuardAccessor(7) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7], 139777662339680) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn, 139777662340880) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k, 139728971383280) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A, 139728971281456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'], 139726887834368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_A['default_2'].weight, 139727353911776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B, 139728971273584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B['default_2'], 139726887836432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.base_layer, 139777662341024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout, 139728971387408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['default_2'], 139726887833696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q, 139728971398544) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A, 139728971396816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'], 139726887831632) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_A['default_2'].weight, 139727353909216) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B, 139728971397728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B['default_2'], 139726887834608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.base_layer, 139777662341120) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout, 139728971397920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['default_2'], 139726887834560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v, 139728971281984) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A, 139728971282608) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'], 139726887833120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_A['default_2'].weight, 139727354197984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B, 139728971283136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B['default_2'], 139726887833552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.base_layer, 139777662341168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout, 139728971279536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['default_2'], 139726887832784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k, 139777662341072) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_k.weight, 139777664181696) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q, 139777662340928) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.norm_q.weight, 139777664579392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.processor, 139777662340832) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm, 139777662340496) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm, 139777662340640) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu, 139777662340544) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear, 139728971250560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A, 139728971235728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'], 139726887837776) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_A['default_2'].weight, 139727353906816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B, 139728971250752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B['default_2'], 139726887837392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.base_layer, 139777662340592) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout, 139728971239376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['default_2'], 139726887837728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp, 139777662340736) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp, 139728971392352) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A, 139728971393696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'], 139726887835808) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_A['default_2'].weight, 139727353920976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B, 139728971393168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B['default_2'], 139726887835760) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.base_layer, 139777662340688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout, 139728971392208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['default_2'], 139726887835040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out, 139728971391584) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[7].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A, 139728971390528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'], 139726887834464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_A['default_2'].weight, 139727353918736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B, 139728971397968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B['default_2'], 139726887834992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.base_layer, 139777662340784) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout, 139728971391920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout['default_2'], 139726887834800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[7].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[7].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[7].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[7].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[7].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[7].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[7]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8], accessed_by=GetItemGuardAccessor(8) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8], 139777662340448) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn, 139777662341648) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k, 139728972261088) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A, 139728972936480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'], 139726887658368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_A['default_2'].weight, 139727354188224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B, 139728972936048) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B['default_2'], 139726887651168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.base_layer, 139777662341792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout, 139728972926688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['default_2'], 139726887662448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q, 139728972251296) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A, 139728972259696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'], 139726887656976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_A['default_2'].weight, 139727354183984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B, 139728972260464) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B['default_2'], 139726887665232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.base_layer, 139777662341888) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout, 139728972251440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['default_2'], 139726887650304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v, 139728972929712) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A, 139728972930288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'], 139726887650832) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_A['default_2'].weight, 139727354196464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B, 139728972936864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B['default_2'], 139726887657840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.base_layer, 139777662341936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout, 139728972938352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['default_2'], 139726887656352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k, 139777662341840) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_k.weight, 139777664477008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q, 139777662341696) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.norm_q.weight, 139777664475008) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.processor, 139777662341600) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm, 139777662341264) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm, 139777662341408) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu, 139777662341312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear, 139728972659152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A, 139728972660640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'], 139726887831680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_A['default_2'].weight, 139727354187024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B, 139728972659584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B['default_2'], 139726887832592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.base_layer, 139777662341360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout, 139728972644464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['default_2'], 139726887832544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp, 139777662341504) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp, 139728972652096) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A, 139728972640800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'], 139726887830720) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_A['default_2'].weight, 139727354195424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B, 139728972632928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B['default_2'], 139726887650928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.base_layer, 139777662341456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout, 139728972646048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['default_2'], 139726887831152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out, 139728972631008) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[8].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A, 139728972632160) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'], 139726887650736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_A['default_2'].weight, 139727354185344) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B, 139728972642336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B['default_2'], 139726887651216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.base_layer, 139777662341552) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout, 139728972643728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout['default_2'], 139726887651840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[8].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[8].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[8].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[8].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[8].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[8].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[8]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9], accessed_by=GetItemGuardAccessor(9) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9], 139777662341216) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn, 139777662342416) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k, 139728971115264) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A, 139728971113536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'], 139726889620512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_A['default_2'].weight, 139727354190384) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B, 139728971106624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B['default_2'], 139726889619792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.base_layer, 139777662342560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout, 139728971117088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['default_2'], 139726889620752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q, 139728971119152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A, 139728971115552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'], 139726889626224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_A['default_2'].weight, 139727354187104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B, 139728971109888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B['default_2'], 139726889619600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.base_layer, 139777662342656) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout, 139728971118672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['default_2'], 139726889625072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v, 139728971106096) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A, 139728971106528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'], 139726889085552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_A['default_2'].weight, 139727354186304) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B, 139728971119248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B['default_2'], 139726889085648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.base_layer, 139777662342704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout, 139728971105856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['default_2'], 139726888856704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k, 139777662342608) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_k.weight, 139777664476128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q, 139777662342464) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.norm_q.weight, 139777664479568) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.processor, 139777662342368) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm, 139777662342032) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm, 139777662342176) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu, 139777662342080) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear, 139728972826368) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A, 139728972833904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'], 139726887650688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_A['default_2'].weight, 139727354190544) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B, 139728972829728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B['default_2'], 139726887649920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.base_layer, 139777662342128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout, 139728972836016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['default_2'], 139726887652944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp, 139777662342272) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp, 139728971175904) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A, 139728971181712) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'], 139726887649536) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_A['default_2'].weight, 139727354195984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B, 139728971185792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B['default_2'], 139726887650352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.base_layer, 139777662342224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout, 139728971184592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['default_2'], 139726887649824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out, 139728971105280) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[9].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A, 139728971117856) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'], 139726887654912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_A['default_2'].weight, 139727354189424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B, 139728971120064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B['default_2'], 139726889622480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.base_layer, 139777662342320) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout, 139728971119104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout['default_2'], 139726887655728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[9].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[9].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[9].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[9].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[9].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[9].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[9]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10], accessed_by=GetItemGuardAccessor(10) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10], 139777662341984) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn, 139777662343184) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k, 139728971470960) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A, 139728971468704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'], 139726889688208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_A['default_2'].weight, 139727356702576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B, 139728971470816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B['default_2'], 139726889687392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.base_layer, 139777662343328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout, 139728971471008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['default_2'], 139726892468928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q, 139728971466352) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A, 139728971471488) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'], 139726892214416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_A['default_2'].weight, 139727358187920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B, 139728971468944) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B['default_2'], 139726892476032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.base_layer, 139777662343424) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout, 139728971466064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['default_2'], 139726892210048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v, 139728971472544) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A, 139728971473216) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'], 139726889871840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_A['default_2'].weight, 139727355474896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B, 139728971473024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B['default_2'], 139726889874960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.base_layer, 139777662343472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout, 139728971472448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['default_2'], 139726889691280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k, 139777662343376) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_k.weight, 139777659800128) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q, 139777662343232) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.norm_q.weight, 139777664329632) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.processor, 139777662343136) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm, 139777662342800) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm, 139777662342944) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu, 139777662342848) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear, 139728971114256) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A, 139728971117040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'], 139726889084880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_A['default_2'].weight, 139727354199504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B, 139728971116176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B['default_2'], 139726889086032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.base_layer, 139777662342896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout, 139728971114736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['default_2'], 139726889083200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp, 139777662343040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp, 139728971106480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A, 139728971106288) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'], 139726889079696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_A['default_2'].weight, 139727354193024) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B, 139728971105808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B['default_2'], 139726889079936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.base_layer, 139777662342992) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout, 139728971106768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['default_2'], 139726889079072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out, 139728971469568) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[10].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A, 139728971469808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'], 139726892325120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_A['default_2'].weight, 139727358186720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B, 139728971470096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B['default_2'], 139726892207648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.base_layer, 139777662343088) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout, 139728971469520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout['default_2'], 139726889083056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[10].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[10].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[10].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[10].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[10].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[10].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[10]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11], accessed_by=GetItemGuardAccessor(11) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11], 139777662342752) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn, 139777662343952) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k, 139728972035072) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A, 139728972026192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'], 139726890154496) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_A['default_2'].weight, 139727353351680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B, 139728972030848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B['default_2'], 139726890154208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.base_layer, 139777662344096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout, 139728972031328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['default_2'], 139726890153968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q, 139728971480368) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A, 139728971480560) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'], 139726890141056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_A['default_2'].weight, 139727353351840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B, 139728971480704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B['default_2'], 139726890150368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.base_layer, 139777662344192) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout, 139728971480272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['default_2'], 139726890406080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v, 139728972022112) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A, 139728972022304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'], 139726890147728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_A['default_2'].weight, 139727353353600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B, 139728971155296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B['default_2'], 139726890147824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.base_layer, 139777662344240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout, 139728972022016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['default_2'], 139726890144128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k, 139777662344144) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_k.weight, 139777660418096) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q, 139777662344000) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.norm_q.weight, 139777659854400) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.processor, 139777662343904) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm, 139777662343568) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm, 139777662343712) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu, 139777662343616) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear, 139728971475952) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A, 139728971476624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'], 139726890406416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_A['default_2'].weight, 139727355462896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B, 139728971477008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B['default_2'], 139726890406992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.base_layer, 139777662343664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout, 139728971473888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['default_2'], 139726890006224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp, 139777662343808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp, 139728971479600) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A, 139728971478256) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'], 139726890405456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_A['default_2'].weight, 139727350217552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B, 139728971479168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B['default_2'], 139726890405792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.base_layer, 139777662343760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout, 139728971479696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['default_2'], 139726890406272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out, 139728971464816) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[11].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A, 139728971478784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'], 139726890401904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_A['default_2'].weight, 139727353351040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B, 139728971478640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B['default_2'], 139726890405120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.base_layer, 139777662343856) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout, 139728971477344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout['default_2'], 139726890404688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[11].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[11].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[11].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[11].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[11].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[11].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[11]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12], accessed_by=GetItemGuardAccessor(12) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12], 139777662343520) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn, 139777662344720) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k, 139728976964640) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A, 139728976963872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'], 139726890303248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_A['default_2'].weight, 139727353658672) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B, 139728976954656) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B['default_2'], 139726890292832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.base_layer, 139777662344864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout, 139728976961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['default_2'], 139726890288704) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q, 139728972691392) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A, 139728976524768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'], 139726890303008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_A['default_2'].weight, 139727357211440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B, 139728976517616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B['default_2'], 139726890298304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.base_layer, 139777662344960) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout, 139728976524384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['default_2'], 139726890298016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v, 139728976959264) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A, 139728976959168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'], 139726890293600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_A['default_2'].weight, 139727353657552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B, 139728976957776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B['default_2'], 139726890300848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.base_layer, 139777662345008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout, 139728976960512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['default_2'], 139726890301904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k, 139777662344912) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_k.weight, 139777664575232) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q, 139777662344768) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.norm_q.weight, 139781198835088) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.processor, 139777662344672) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm, 139777662344336) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm, 139777662344480) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu, 139777662344384) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear, 139728971163312) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A, 139728972861584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'], 139726890300656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_A['default_2'].weight, 139727356198512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B, 139728972869984) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B['default_2'], 139726890290720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.base_layer, 139777662344432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout, 139728971169072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['default_2'], 139726890155600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp, 139777662344576) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp, 139728972862976) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A, 139728978267248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'], 139726890287888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_A['default_2'].weight, 139727356194592) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B, 139728978270704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B['default_2'], 139726890297728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.base_layer, 139777662344528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout, 139728972863120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['default_2'], 139726890287504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out, 139728978280400) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[12].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A, 139728978277904) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'], 139726890302288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_A['default_2'].weight, 139727357204880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B, 139728978268976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B['default_2'], 139726890298352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.base_layer, 139777662344624) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout, 139728978274736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout['default_2'], 139726890297824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[12].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[12].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[12].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[12].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[12].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[12].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[12]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13], accessed_by=GetItemGuardAccessor(13) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13], 139777662344288) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn, 139777662345488) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k, 139728976676128) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A, 139728976687312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'], 139726891639344) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_A['default_2'].weight, 139727356821424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B, 139728976682896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B['default_2'], 139726891639536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.base_layer, 139777662345632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout, 139728976675408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['default_2'], 139726891638960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q, 139728978261136) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A, 139728978256720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'], 139726891639248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_A['default_2'].weight, 139727358932400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B, 139728978254032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B['default_2'], 139726891640112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.base_layer, 139777662345728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout, 139728978258016) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['default_2'], 139726891639680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v, 139728977839920) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A, 139728977850768) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'], 139726891638432) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_A['default_2'].weight, 139727356815184) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B, 139728977849424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B['default_2'], 139726891639056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.base_layer, 139777662345776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout, 139728977850432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['default_2'], 139726891638672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k, 139777662345680) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_k.weight, 139777664581312) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q, 139777662345536) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.norm_q.weight, 139777659801808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.processor, 139777662345440) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm, 139777662345104) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm, 139777662345248) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu, 139777662345152) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear, 139728976958160) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A, 139728976958640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'], 139726891641888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_A['default_2'].weight, 139727353659232) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B, 139728976956144) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B['default_2'], 139726891643232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.base_layer, 139777662345200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout, 139728976957248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['default_2'], 139726890300128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp, 139777662345344) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp, 139728976968096) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A, 139728976967808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'], 139726891641072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_A['default_2'].weight, 139727354917840) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B, 139728976968528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B['default_2'], 139726891641648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.base_layer, 139777662345296) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout, 139728976953600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['default_2'], 139726891641360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out, 139728976969152) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[13].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A, 139728976965744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'], 139726891640736) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_A['default_2'].weight, 139727359055552) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B, 139728978254800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B['default_2'], 139726891640304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.base_layer, 139777662345392) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout, 139728976954416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout['default_2'], 139726891640592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[13].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[13].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[13].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[13].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[13].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[13].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[13]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14], accessed_by=GetItemGuardAccessor(14) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14], 139777662345056) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn, 139777662346256) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k, 139728976866912) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A, 139728976857312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'], 139726891635120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_A['default_2'].weight, 139727350312096) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B, 139728976861920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B['default_2'], 139726891634544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.base_layer, 139777662346400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout, 139728976866528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['default_2'], 139726891635072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q, 139728977057712) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A, 139728977062128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'], 139726891635552) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_A['default_2'].weight, 139727359441424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B, 139728977064912) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B['default_2'], 139726891636176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.base_layer, 139777662346496) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout, 139728977052288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['default_2'], 139726891635792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v, 139728976856304) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A, 139728976856928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'], 139726891633488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_A['default_2'].weight, 139727359244176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B, 139728976855824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B['default_2'], 139726891634400) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.base_layer, 139777662346544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout, 139728976858080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['default_2'], 139726891634496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k, 139777662346448) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_k.weight, 139777664751776) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q, 139777662346304) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.norm_q.weight, 139777664573952) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.processor, 139777662346208) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm, 139777662345872) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm, 139777662346016) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu, 139777662345920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear, 139728978363056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A, 139728978365984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'], 139726891638000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_A['default_2'].weight, 139727354428640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B, 139728978376064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B['default_2'], 139726891637424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.base_layer, 139777662345968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout, 139728978373424) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['default_2'], 139726891637952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp, 139777662346112) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp, 139728978369248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A, 139728978363008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'], 139726891636320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_A['default_2'].weight, 139727356602192) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B, 139728978371024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B['default_2'], 139726891637280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.base_layer, 139777662346064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout, 139728978373568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['default_2'], 139726891637376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out, 139728978271616) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[14].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A, 139728977064672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'], 139726891636464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_A['default_2'].weight, 139727354284928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B, 139728977054832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B['default_2'], 139726891636560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.base_layer, 139777662346160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout, 139728977052864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout['default_2'], 139726891636080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[14].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[14].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[14].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[14].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[14].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[14].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[14]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15], accessed_by=GetItemGuardAccessor(15) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15], 139777662345824) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn, 139777662347024) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k, 139728976897328) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A, 139728976891760) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'], 139726891495392) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_A['default_2'].weight, 139727357867600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B, 139728976895024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B['default_2'], 139726891494672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.base_layer, 139777662347168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout, 139728976897280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['default_2'], 139726891630896) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q, 139728976717392) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A, 139728976716528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'], 139726891630800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_A['default_2'].weight, 139727355049312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B, 139728976717584) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B['default_2'], 139726891631088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.base_layer, 139777662347264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout, 139728976719936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['default_2'], 139726891631232) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v, 139728978363584) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A, 139728977147120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'], 139726891499184) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_A['default_2'].weight, 139727358150032) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B, 139728977145968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B['default_2'], 139726891498752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.base_layer, 139777662347312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout, 139728977147264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['default_2'], 139726891498992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k, 139777662347216) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_k.weight, 139777659855920) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q, 139777662347072) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.norm_q.weight, 139777660118384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.processor, 139777662346976) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm, 139777662346640) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm, 139777662346784) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu, 139777662346688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear, 139728976862928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A, 139728976867056) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'], 139726891633584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_A['default_2'].weight, 139727358249776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B, 139728976862016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B['default_2'], 139726891633632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.base_layer, 139777662346736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout, 139728976866816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['default_2'], 139726891633200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp, 139777662346880) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp, 139728976866864) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A, 139728976855440) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'], 139726891633008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_A['default_2'].weight, 139727360341744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B, 139728976759488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B['default_2'], 139726891632528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.base_layer, 139777662346832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout, 139728976856256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['default_2'], 139726891632960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out, 139728976761504) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[15].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A, 139728976765152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'], 139726891632336) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_A['default_2'].weight, 139727354604704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B, 139728976760208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B['default_2'], 139726891631952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.base_layer, 139777662346928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout, 139728976772832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout['default_2'], 139726891632288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[15].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[15].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[15].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[15].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[15].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[15].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[15]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16], accessed_by=GetItemGuardAccessor(16) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16], 139777662346592) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn, 139777662347792) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k, 139728977139200) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A, 139728977138624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'], 139726891496928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_A['default_2'].weight, 139727355767488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B, 139728977146208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B['default_2'], 139726891489488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.base_layer, 139777662347936) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout, 139728977140352) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['default_2'], 139726891489200) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q, 139728977143232) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A, 139728977145056) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'], 139726891489056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_A['default_2'].weight, 139727358602160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B, 139728977142320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B['default_2'], 139726891489104) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.base_layer, 139777662348032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout, 139728977140880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['default_2'], 139726891489680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v, 139728977090576) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A, 139728977091632) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'], 139726891496352) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_A['default_2'].weight, 139727355771728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B, 139728977097056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B['default_2'], 139726891495968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.base_layer, 139777662348080) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout, 139728977092496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['default_2'], 139726891496304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k, 139777662347984) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_k.weight, 139777664323712) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q, 139777662347840) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.norm_q.weight, 139777664746576) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.processor, 139777662347744) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm, 139777662347408) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm, 139777662347552) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu, 139777662347456) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear, 139728977148032) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A, 139728977147936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'], 139726891498512) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_A['default_2'].weight, 139727354201168) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B, 139728977140784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B['default_2'], 139726891498704) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.base_layer, 139777662347504) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout, 139728977139248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['default_2'], 139726891498464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp, 139777662347648) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp, 139728977146544) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A, 139728977144624) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'], 139726891497600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_A['default_2'].weight, 139727357307984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B, 139728977146064) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B['default_2'], 139726891497120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.base_layer, 139777662347600) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout, 139728977148848) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['default_2'], 139726891497408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out, 139728977141264) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[16].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A, 139728977139680) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'], 139726891488864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_A['default_2'].weight, 139727357273536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B, 139728977139488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B['default_2'], 139726891487616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.base_layer, 139777662347696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout, 139728977137904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout['default_2'], 139726891485408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[16].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[16].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[16].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[16].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[16].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[16].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[16]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17], accessed_by=GetItemGuardAccessor(17) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17], 139777662347360) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn, 139777662348560) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k, 139728976902608) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A, 139728984827088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'], 139726891491696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_A['default_2'].weight, 139727349090560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B, 139728984821040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B['default_2'], 139726891492128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.base_layer, 139777662348704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout, 139728984821952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['default_2'], 139726891491360) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q, 139728977241440) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A, 139728977241584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'], 139726891492944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_A['default_2'].weight, 139727356087328) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B, 139728977247056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B['default_2'], 139726891495008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.base_layer, 139777662348800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout, 139728977239856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['default_2'], 139726891492272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v, 139728984951008) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A, 139728984950480) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'], 139726891490256) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_A['default_2'].weight, 139727349098240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B, 139728984961616) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B['default_2'], 139726891491168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.base_layer, 139777662348848) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout, 139728984962528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['default_2'], 139726891491120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k, 139777662348752) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_k.weight, 139777660124544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q, 139777662348608) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.norm_q.weight, 139777659810208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.processor, 139777662348512) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm, 139777662348176) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm, 139777662348320) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu, 139777662348224) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear, 139728977091920) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A, 139728977089664) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'], 139726891494384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_A['default_2'].weight, 139727355567120) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B, 139728977089856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B['default_2'], 139726891494336) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.base_layer, 139777662348272) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout, 139728977086112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['default_2'], 139726891493616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp, 139777662348416) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp, 139728977092400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A, 139728977089136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'], 139726891493040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_A['default_2'].weight, 139727355563040) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B, 139728977093696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B['default_2'], 139726891493568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.base_layer, 139777662348368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout, 139728977096960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['default_2'], 139726891493376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out, 139728977238848) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[17].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A, 139728977241008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'], 139726891490208) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_A['default_2'].weight, 139727357763856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B, 139728977238272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B['default_2'], 139726891493184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.base_layer, 139777662348464) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout, 139728977238944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout['default_2'], 139726891493136) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[17].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[17].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[17].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[17].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[17].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[17].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[17]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18], accessed_by=GetItemGuardAccessor(18) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18], 139777662348128) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn, 139777662349328) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k, 139728977232752) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A, 139728985170096) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'], 139726891375376) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_A['default_2'].weight, 139727353392528) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B, 139728985172784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B['default_2'], 139726891381952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.base_layer, 139777662349472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout, 139728985057808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['default_2'], 139726891380464) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q, 139728985001600) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A, 139728985003952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'], 139726891380320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_A['default_2'].weight, 139727353393648) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B, 139728985013456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B['default_2'], 139726891382384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.base_layer, 139777662349568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout, 139728985011248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['default_2'], 139726891379840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v, 139728985237456) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A, 139728985237648) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'], 139726891374128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_A['default_2'].weight, 139727353383488) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B, 139728985228096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B['default_2'], 139726891373936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.base_layer, 139777662349616) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout, 139728985239040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['default_2'], 139726891374800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k, 139777662349520) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_k.weight, 139777660373728) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q, 139777662349376) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.norm_q.weight, 139777660372208) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.processor, 139777662349280) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm, 139777662348944) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm, 139777662349088) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu, 139777662348992) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear, 139728984176528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A, 139728984167408) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'], 139726891486944) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_A['default_2'].weight, 139727349098080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B, 139728984164576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B['default_2'], 139726891489920) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.base_layer, 139777662349040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout, 139728984170336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['default_2'], 139726891489296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp, 139777662349184) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp, 139728985421472) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A, 139728985413264) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'], 139726891375520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_A['default_2'].weight, 139727349095200) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B, 139728985412448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B['default_2'], 139726891377392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.base_layer, 139777662349136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout, 139728985420512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['default_2'], 139726891378592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out, 139728985413552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[18].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A, 139728985412016) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'], 139726891372976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_A['default_2'].weight, 139727349096080) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B, 139728985421232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B['default_2'], 139726891379120) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.base_layer, 139777662349232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout, 139728985409040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout['default_2'], 139726891381088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[18].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[18].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[18].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[18].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[18].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[18].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[18]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19], accessed_by=GetItemGuardAccessor(19) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19], 139777662348896) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn, 139777662350096) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k, 139728983778512) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A, 139728983777168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'], 139726890767680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_A['default_2'].weight, 139727358314912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B, 139728983772560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B['default_2'], 139726890768592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.base_layer, 139777662350240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout, 139728983776688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['default_2'], 139726890772720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q, 139728985529056) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A, 139728985538224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'], 139726891247136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_A['default_2'].weight, 139727355885936) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B, 139728983785088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B['default_2'], 139726891247712) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.base_layer, 139777662350336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout, 139728985534384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['default_2'], 139726892705264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v, 139728985587472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A, 139728985577920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'], 139726890775456) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_A['default_2'].weight, 139727358818912) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B, 139728985571776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B['default_2'], 139726890775552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.base_layer, 139777662350384) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout, 139728985584640) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['default_2'], 139726890776752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k, 139777662350288) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_k.weight, 139777664712304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q, 139777662350144) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.norm_q.weight, 139777660377488) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.processor, 139777662350048) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm, 139777662349712) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm, 139777662349856) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu, 139777662349760) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear, 139728985268016) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A, 139728985267248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'], 139726891373696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_A['default_2'].weight, 139727357117920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B, 139728985271568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B['default_2'], 139726891373744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.base_layer, 139777662349808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout, 139728985264512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['default_2'], 139726891379984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp, 139777662349952) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp, 139728985378528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A, 139728985482976) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'], 139726891379504) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_A['default_2'].weight, 139727357070928) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B, 139728985483936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B['default_2'], 139726891249440) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.base_layer, 139777662349904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout, 139728985390288) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['default_2'], 139726891369376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out, 139728985473616) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[19].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A, 139728985479184) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'], 139726891249680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_A['default_2'].weight, 139727356415088) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B, 139728985531264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B['default_2'], 139726891245024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.base_layer, 139777662350000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout, 139728985482832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout['default_2'], 139726891250112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[19].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[19].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[19].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[19].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[19].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[19].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[19]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20], accessed_by=GetItemGuardAccessor(20) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20], 139777662349664) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn, 139777662350864) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k, 139728985675840) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A, 139728985674784) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'], 139726893850224) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_A['default_2'].weight, 139727359355744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B, 139728985681696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B['default_2'], 139726893857040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.base_layer, 139777662351008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout, 139728985676128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['default_2'], 139726893856128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q, 139728985775152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A, 139728985772128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'], 139726893855408) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_A['default_2'].weight, 139727350504944) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B, 139728985783840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B['default_2'], 139726893856896) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.base_layer, 139777662351104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout, 139728985776448) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['default_2'], 139726893353376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v, 139728985014224) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A, 139729027839504) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'], 139726893853296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_A['default_2'].weight, 139727359107984) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B, 139729027836960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B['default_2'], 139726893856608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.base_layer, 139777662351152) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout, 139729027839168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['default_2'], 139726893853584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k, 139777662351056) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_k.weight, 139777664182496) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q, 139777662350912) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.norm_q.weight, 139777664729392) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.processor, 139777662350816) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm, 139777662350480) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm, 139777662350624) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu, 139777662350528) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear, 139728984115040) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A, 139728984117104) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'], 139726895931616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_A['default_2'].weight, 139727356137520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B, 139728984115184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B['default_2'], 139726895924320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.base_layer, 139777662350576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout, 139728984118496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['default_2'], 139726893947808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp, 139777662350720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp, 139728984121712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A, 139728984121424) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'], 139726895936704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_A['default_2'].weight, 139727354669520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B, 139728984127856) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B['default_2'], 139726895937088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.base_layer, 139777662350672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout, 139728984123440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['default_2'], 139726895936128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out, 139728985725904) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[20].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A, 139728985726672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'], 139726893362688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_A['default_2'].weight, 139727354661600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B, 139728985730080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B['default_2'], 139726893362496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.base_layer, 139777662350768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout, 139728985725760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout['default_2'], 139726893352176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[20].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[20].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[20].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[20].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[20].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[20].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[20]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21], accessed_by=GetItemGuardAccessor(21) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21], 139777662350432) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn, 139777662351632) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k, 139729028935408) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A, 139729071121568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'], 139726891917728) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_A['default_2'].weight, 139727353427536) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B, 139729071115664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B['default_2'], 139726891917008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.base_layer, 139777662351776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout, 139729071261584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['default_2'], 139726891920560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q, 139729027941952) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A, 139729070046176) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'], 139726891915856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_A['default_2'].weight, 139727353426576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B, 139729071757088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B['default_2'], 139726891920176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.base_layer, 139777662351872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout, 139729070400864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['default_2'], 139726891919936) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v, 139729071458816) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A, 139729073486720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'], 139726892017616) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_A['default_2'].weight, 139727353414176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B, 139729073485520) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B['default_2'], 139726892012672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.base_layer, 139777662351920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout, 139729072770544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['default_2'], 139726892017568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k, 139777662351824) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_k.weight, 139777660377808) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q, 139777662351680) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.norm_q.weight, 139777664725328) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.processor, 139777662351584) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm, 139777662351248) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm, 139777662351392) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu, 139777662351296) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear, 139729031521440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A, 139729030869728) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'], 139726891912304) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_A['default_2'].weight, 139727355320000) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B, 139729073071840) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B['default_2'], 139726891914560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.base_layer, 139777662351344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout, 139729030517024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['default_2'], 139726893855792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp, 139777662351488) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp, 139729029447152) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A, 139729029435152) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'], 139726891919840) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_A['default_2'].weight, 139727353118240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B, 139729029300480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B['default_2'], 139726891923680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.base_layer, 139777662351440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout, 139729029446864) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['default_2'], 139726891922576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out, 139729029021120) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[21].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A, 139729029123808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'], 139726891925312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_A['default_2'].weight, 139727351142480) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B, 139729029121936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B['default_2'], 139726891923728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.base_layer, 139777662351536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout, 139729029132256) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout['default_2'], 139726891925264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[21].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[21].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[21].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[21].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[21].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[21].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[21]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22], accessed_by=GetItemGuardAccessor(22) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22], 139777662351200) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn, 139777662352400) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k, 139729075138144) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A, 139729075128688) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'], 139726893284176) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_A['default_2'].weight, 139727356875520) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B, 139729076787024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B['default_2'], 139726893285088) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.base_layer, 139777662352544) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout, 139729075138480) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['default_2'], 139726893284992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q, 139729073604208) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A, 139729075662192) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'], 139726893269872) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_A['default_2'].weight, 139727356873920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B, 139729075661328) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B['default_2'], 139726893269392) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.base_layer, 139777662352640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout, 139729075660608) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['default_2'], 139726892010224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v, 139729078054256) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A, 139729078057568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'], 139726893284032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_A['default_2'].weight, 139727353258816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B, 139729078050320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B['default_2'], 139726893284128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.base_layer, 139777662352688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout, 139729078055120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['default_2'], 139726893283648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k, 139777662352592) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_k.weight, 139777664757856) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q, 139777662352448) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.norm_q.weight, 139777660378528) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.processor, 139777662352352) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm, 139777662352016) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm, 139777662352160) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu, 139777662352064) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear, 139729072306688) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A, 139729072306448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'], 139726892016560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_A['default_2'].weight, 139727353416416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B, 139729072306208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B['default_2'], 139726892016224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.base_layer, 139777662352112) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout, 139729072294208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['default_2'], 139726892016368) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp, 139777662352256) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp, 139729072038160) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A, 139729071776400) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'], 139726892015120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_A['default_2'].weight, 139727353414736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B, 139729071770016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B['default_2'], 139726892013968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.base_layer, 139777662352208) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout, 139729071783648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['default_2'], 139726892015696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out, 139729070823680) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[22].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A, 139729073304432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'], 139726892009888) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_A['default_2'].weight, 139727356882800) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B, 139729073304528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B['default_2'], 139726892010272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.base_layer, 139777662352304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout, 139729073299296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout['default_2'], 139726892009024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[22].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[22].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[22].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[22].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[22].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[22].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[22]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23], accessed_by=GetItemGuardAccessor(23) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23], 139777662351968) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn, 139777662353168) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k, 139729076457856) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A, 139729076456224) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'], 139726893280672) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_A['default_2'].weight, 139727354571616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B, 139729076457232) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B['default_2'], 139726893281200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.base_layer, 139777661026368) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout, 139729076466736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['default_2'], 139726893280624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q, 139729074952976) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A, 139729074950432) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'], 139726893281488) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_A['default_2'].weight, 139727353260576) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B, 139729074946496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B['default_2'], 139726893281872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.base_layer, 139777662353312) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout, 139729074961664) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['default_2'], 139726893280768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v, 139729076356960) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A, 139729077922128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'], 139726893279856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_A['default_2'].weight, 139727355814160) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B, 139729077927504) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B['default_2'], 139726893279472) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.base_layer, 139777661026416) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout, 139729077912144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['default_2'], 139726893279904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k, 139777662353360) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_k.weight, 139777660445824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q, 139777662353216) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.norm_q.weight, 139777664332992) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.processor, 139777662353120) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm, 139777662352784) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm, 139777662352928) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu, 139777662352832) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear, 139728986592768) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A, 139728986129744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'], 139726893283264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_A['default_2'].weight, 139727353256736) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B, 139728986142176) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B['default_2'], 139726893283792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.base_layer, 139777662352880) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout, 139728986590560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['default_2'], 139726893283456) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp, 139777662353024) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp, 139728986134400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A, 139728986136272) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'], 139726893282928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_A['default_2'].weight, 139727353252976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B, 139728986134736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B['default_2'], 139726893282352) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.base_layer, 139777662352976) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout, 139728986133104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['default_2'], 139726893282880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out, 139728986503344) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[23].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A, 139728986498544) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'], 139726893281968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_A['default_2'].weight, 139727353266016) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B, 139729076928080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B['default_2'], 139726893282448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.base_layer, 139777662353072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout, 139728986495088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout['default_2'], 139726893282400) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[23].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[23].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[23].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[23].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[23].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[23].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[23]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24], accessed_by=GetItemGuardAccessor(24) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24], 139777662352736) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn, 139777661026896) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k, 139729080562160) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A, 139729080563120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'], 139726893276112) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_A['default_2'].weight, 139727354111584) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B, 139729080563600) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B['default_2'], 139726893276496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.base_layer, 139777661027040) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout, 139729080562592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['default_2'], 139726893275872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q, 139729079640528) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A, 139729079641824) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'], 139726893277120) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_A['default_2'].weight, 139727354111904) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B, 139729079632080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B['default_2'], 139726893276784) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.base_layer, 139777661027136) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout, 139729079637408) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['default_2'], 139726893276976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v, 139729080553184) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A, 139729080557936) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'], 139726893275584) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_A['default_2'].weight, 139727354104864) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B, 139729080552992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B['default_2'], 139726893275776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.base_layer, 139777661027184) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout, 139729080562496) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['default_2'], 139726893275296) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k, 139777661027088) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_k.weight, 139777660322976) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q, 139777661026944) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.norm_q.weight, 139777664578432) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.processor, 139777661026848) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm, 139777661026512) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm, 139777661026656) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu, 139777661026560) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear, 139729077811472) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A, 139729077580800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'], 139726893279088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_A['default_2'].weight, 139727357446880) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B, 139729077576864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B['default_2'], 139726893279568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.base_layer, 139777661026608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout, 139729077812960) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['default_2'], 139726893279520) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp, 139777661026752) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp, 139729079646912) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A, 139729079643840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'], 139726893278608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_A['default_2'].weight, 139727357456320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B, 139729079638608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B['default_2'], 139726893278992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.base_layer, 139777661026704) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout, 139729079647344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['default_2'], 139726893277888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out, 139729079643936) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[24].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A, 139729079634960) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'], 139726893277792) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_A['default_2'].weight, 139727354108784) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B, 139729079635728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B['default_2'], 139726893278320) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.base_layer, 139777661026800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout, 139729079646384) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout['default_2'], 139726893277744) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[24].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[24].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[24].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[24].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[24].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[24].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[24]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25], accessed_by=GetItemGuardAccessor(25) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25], 139777661026464) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn, 139777661027664) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k, 139729081832768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A, 139729081831280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'], 139726893271600) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_A['default_2'].weight, 139727354105264) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B, 139729081840016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B['default_2'], 139726893272080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.base_layer, 139777661027808) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout, 139729081828112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['default_2'], 139726893271696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q, 139729081832816) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A, 139729081828352) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'], 139726893272464) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_A['default_2'].weight, 139727354112704) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B, 139729081828640) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B['default_2'], 139726893272992) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.base_layer, 139777661027904) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout, 139729081835792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['default_2'], 139726893272176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v, 139729082854144) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A, 139729082857168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'], 139726893271264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_A['default_2'].weight, 139727354105104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B, 139729082857936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B['default_2'], 139726893270880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.base_layer, 139777661027952) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout, 139729082855536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['default_2'], 139726893271024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k, 139777661027856) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_k.weight, 139777664182896) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q, 139777661027712) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.norm_q.weight, 139777664566032) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.processor, 139777661027616) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm, 139777661027280) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm, 139777661027424) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu, 139777661027328) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear, 139729080550784) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A, 139729080552368) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'], 139726893274624) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_A['default_2'].weight, 139727354116624) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B, 139729080550736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B['default_2'], 139726893275200) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.base_layer, 139777661027376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout, 139729080549728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['default_2'], 139726893274432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp, 139777661027520) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp, 139729080556496) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A, 139729081837952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'], 139726893274240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_A['default_2'].weight, 139727354108464) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B, 139729081834208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B['default_2'], 139726893273904) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.base_layer, 139777661027472) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout, 139729080553328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['default_2'], 139726893273952) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out, 139729081843424) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[25].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A, 139729081839584) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'], 139726893272608) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_A['default_2'].weight, 139727354115104) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B, 139729081837040) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B['default_2'], 139726893273664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.base_layer, 139777661027568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout, 139729081840832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout['default_2'], 139726893273760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[25].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[25].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[25].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[25].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[25].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[25].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[25]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26], accessed_by=GetItemGuardAccessor(26) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26], 139777661027232) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn, 139777661028432) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k, 139729085484240) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A, 139729085485200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'], 139726893136912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_A['default_2'].weight, 139727354110144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B, 139729085485680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B['default_2'], 139726893137488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.base_layer, 139777661028576) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout, 139729085484672) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['default_2'], 139726893137440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q, 139729085484576) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A, 139729085491008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'], 139726893137680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_A['default_2'].weight, 139727354117664) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B, 139729085491488) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B['default_2'], 139726893134368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.base_layer, 139777661028672) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout, 139729085493600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['default_2'], 139726893132736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v, 139729085490720) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A, 139729085481072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'], 139726893136528) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_A['default_2'].weight, 139727354113744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B, 139729085490288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B['default_2'], 139726893136192) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.base_layer, 139777661028720) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout, 139729085490624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['default_2'], 139726893136336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k, 139777661028624) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_k.weight, 139777661092384) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q, 139777661028480) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.norm_q.weight, 139777661093744) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.processor, 139777661028384) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm, 139777661028048) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm, 139777661028192) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu, 139777661028096) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear, 139729082856928) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A, 139729082852752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'], 139726893269104) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_A['default_2'].weight, 139727354115744) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B, 139729082850496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B['default_2'], 139726893270736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.base_layer, 139777661028144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout, 139729082857264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['default_2'], 139726893270832) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp, 139777661028288) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp, 139729082848528) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A, 139729082843872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'], 139726893136768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_A['default_2'].weight, 139727354113504) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B, 139729082844112) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B['default_2'], 139726893137824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.base_layer, 139777661028240) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout, 139729082850304) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['default_2'], 139726893269248) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out, 139729082848720) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[26].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A, 139729082857840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'], 139726893123088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_A['default_2'].weight, 139727354106064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B, 139729085496000) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B['default_2'], 139726893131536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.base_layer, 139777661028336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout, 139729082855920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout['default_2'], 139726893126784) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[26].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[26].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[26].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[26].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[26].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[26].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[26]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27], accessed_by=GetItemGuardAccessor(27) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27], 139777661028000) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn, 139777661029200) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k, 139729087700448) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A, 139729087699008) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'], 139726893127312) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_A['default_2'].weight, 139727352152208) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B, 139729087699296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B['default_2'], 139726893128224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.base_layer, 139777661029344) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout, 139729087704720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['default_2'], 139726893124096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q, 139729073171488) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A, 139729087694208) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'], 139726893133072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_A['default_2'].weight, 139727354111424) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B, 139729087698864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B['default_2'], 139726893133648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.base_layer, 139777661029440) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout, 139729087703568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['default_2'], 139726893133264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v, 139729087693392) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A, 139729087693536) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'], 139726893010320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_A['default_2'].weight, 139727352160448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B, 139729087693296) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B['default_2'], 139726893007536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.base_layer, 139777661029488) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout, 139729087693008) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['default_2'], 139726893131104) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k, 139777661029392) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_k.weight, 139777661093504) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q, 139777661029248) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.norm_q.weight, 139777661094864) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.processor, 139777661029152) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm, 139777661028816) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm, 139777661028960) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu, 139777661028864) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear, 139729086476304) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A, 139729086478752) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'], 139726893134416) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_A['default_2'].weight, 139727354116144) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B, 139729086478608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B['default_2'], 139726893133168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.base_layer, 139777661028912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout, 139729086477696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['default_2'], 139726893133216) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp, 139777661029056) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp, 139729086469584) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A, 139729086471840) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'], 139726893135232) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_A['default_2'].weight, 139727354112064) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B, 139729086466560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B['default_2'], 139726893135424) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.base_layer, 139777661029008) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout, 139729086472224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['default_2'], 139726893135472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out, 139729086466416) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[27].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A, 139729086468816) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'], 139726893134560) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_A['default_2'].weight, 139727354116224) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B, 139729086469248) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B['default_2'], 139726893134800) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.base_layer, 139777661029104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout, 139729086464112) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout['default_2'], 139726893134512) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[27].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[27].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[27].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[27].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[27].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[27].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[27]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28], accessed_by=GetItemGuardAccessor(28) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28], 139777661028768) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn, 139777661030016) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k, 139729091443536) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A, 139729092148384) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'], 139726893017040) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_A['default_2'].weight, 139727352166128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B, 139729092139648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B['default_2'], 139726893017664) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.base_layer, 139777661030160) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout, 139729092142528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['default_2'], 139726893014688) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q, 139729091429616) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A, 139729091429280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'], 139726893019056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_A['default_2'].weight, 139727352155808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B, 139729091431680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B['default_2'], 139726893009072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.base_layer, 139777661030256) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout, 139729091429712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['default_2'], 139726893019776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v, 139729092148624) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A, 139729092140992) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'], 139726892906384) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_A['default_2'].weight, 139727352166448) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B, 139729092137152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B['default_2'], 139726892900432) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.base_layer, 139777661030304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout, 139729092143968) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['default_2'], 139726893011280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k, 139777661030208) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_k.weight, 139777661094624) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q, 139777661030064) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.norm_q.weight, 139777661096064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.processor, 139777661029968) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm, 139777661029584) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm, 139777661029728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu, 139777661029632) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear, 139729087705872) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A, 139729087707792) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'], 139726893019296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_A['default_2'].weight, 139727352154688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B, 139729087708272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B['default_2'], 139726893012240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.base_layer, 139777661029680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout, 139729087695792) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['default_2'], 139726893008544) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp, 139777661029872) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp, 139729091442720) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A, 139729091439984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'], 139726893014976) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_A['default_2'].weight, 139727352159968) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B, 139729091439888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B['default_2'], 139726893018864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.base_layer, 139777661029776) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout, 139729091442768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['default_2'], 139726893015312) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out, 139729091434944) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[28].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A, 139729091435040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'], 139726893019680) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_A['default_2'].weight, 139727352164768) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B, 139729091434368) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B['default_2'], 139726893017280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.base_layer, 139777661029920) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout, 139729091437056) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout['default_2'], 139726893022224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[28].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[28].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[28].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[28].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[28].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[28].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[28]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29], accessed_by=GetItemGuardAccessor(29) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29], 139777661029536) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn, 139777661030784) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k, 139729093490768) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A, 139729093490144) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'], 139726895320032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_A['default_2'].weight, 139727352154048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B, 139729093486544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B['default_2'], 139727233341888) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.base_layer, 139777661030928) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout, 139729093483760) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['default_2'], 139727233188048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q, 139729093484624) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A, 139729093491872) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'], 139727232049856) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_A['default_2'].weight, 139727352156288) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B, 139729093484768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B['default_2'], 139727233188720) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.base_layer, 139777661031024) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout, 139729093482800) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['default_2'], 139726894565168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v, 139729093483472) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A, 139729093484576) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'], 139726895486320) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_A['default_2'].weight, 139727352163888) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B, 139729093481744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B['default_2'], 139726895487136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.base_layer, 139777661031072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout, 139729093481984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['default_2'], 139726895493808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k, 139777661030976) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_k.weight, 139777661095824) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q, 139777661030832) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.norm_q.weight, 139777661097184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.processor, 139777661030736) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm, 139777661030400) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm, 139777661030544) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu, 139777661030448) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear, 139729092135568) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A, 139729092139744) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'], 139726892907440) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_A['default_2'].weight, 139727352159568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B, 139729092141376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B['default_2'], 139726892905808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.base_layer, 139777661030496) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout, 139729092133264) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['default_2'], 139726892899472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp, 139777661030640) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp, 139729092147040) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A, 139729092147280) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'], 139726894440768) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_A['default_2'].weight, 139727352160688) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B, 139729092137824) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B['default_2'], 139726894441536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.base_layer, 139777661030592) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout, 139729092141808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['default_2'], 139726894311376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out, 139729092133552) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[29].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A, 139729092134800) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'], 139726894565264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_A['default_2'].weight, 139727352153808) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B, 139729092132928) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B['default_2'], 139726894564496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.base_layer, 139777661030688) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout, 139729092133648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout['default_2'], 139726894440144) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[29].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[29].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[29].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[29].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[29].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[29].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[29]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30], accessed_by=GetItemGuardAccessor(30) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30], 139777661030352) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn, 139777661031552) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k, 139729091698624) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A, 139729091700928) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'], 139726895661008) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_A['default_2'].weight, 139727352159728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B, 139729091692672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B['default_2'], 139726895650976) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.base_layer, 139777661031696) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout, 139729091697376) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['default_2'], 139726895658128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q, 139729091695024) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A, 139729091697232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'], 139726895647712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_A['default_2'].weight, 139727352165568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B, 139729091696752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B['default_2'], 139726895646128) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.base_layer, 139777661031792) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout, 139729091695120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['default_2'], 139726895659616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v, 139729093047056) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A, 139729093043168) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'], 139726895758064) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_A['default_2'].weight, 139727352167728) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B, 139729093043072) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B['default_2'], 139726895753744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.base_layer, 139777661031840) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout, 139729093047680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['default_2'], 139726895651168) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k, 139777661031744) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_k.weight, 139777661096944) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q, 139777661031600) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.norm_q.weight, 139777661098304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.processor, 139777661031504) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm, 139777661031168) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm, 139777661031312) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu, 139777661031216) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear, 139729093480016) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A, 139729093480976) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'], 139726895496928) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_A['default_2'].weight, 139727352168128) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B, 139729093481024) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B['default_2'], 139726895649680) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.base_layer, 139777661031264) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout, 139729093480064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['default_2'], 139726895491984) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp, 139777661031408) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp, 139729091705296) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A, 139729091703232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'], 139726895652656) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_A['default_2'].weight, 139727352168048) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B, 139729091699968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B['default_2'], 139726895644880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.base_layer, 139777661031360) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout, 139729091706736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['default_2'], 139726895652272) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out, 139729091701456) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[30].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A, 139729091691472) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'], 139726895659520) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_A['default_2'].weight, 139727352163568) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B, 139729091690848) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B['default_2'], 139726895648816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.base_layer, 139777661031456) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout, 139729091700592) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout['default_2'], 139726895659328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[30].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[30].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[30].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[30].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[30].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[30].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[30]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31], accessed_by=GetItemGuardAccessor(31) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31], 139777661031120) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn, 139777661032320) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k, 139729093175344) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A, 139729093165984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'], 139726894994128) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_A['default_2'].weight, 139727353835696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B, 139729093170736) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B['default_2'], 139726895002096) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.base_layer, 139777661032464) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout, 139729093166224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['default_2'], 139726895743040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q, 139729093177696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A, 139729093178656) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'], 139726895748800) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_A['default_2'].weight, 139727353823616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B, 139729093179136) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B['default_2'], 139726895757152) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.base_layer, 139777661032560) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout, 139729093178128) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['default_2'], 139726895748080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v, 139729093173328) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A, 139729093165696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'], 139726894993648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_A['default_2'].weight, 139727353834896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B, 139729093170448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B['default_2'], 139726894993696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.base_layer, 139777661032608) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout, 139729093172944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['default_2'], 139726894993120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k, 139777661032512) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_k.weight, 139777661098064) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q, 139777661032368) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.norm_q.weight, 139777661099424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.processor, 139777661032272) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm, 139777661031936) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm, 139777661032080) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu, 139777661031984) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear, 139729093035440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A, 139729093035200) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'], 139726895756864) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_A['default_2'].weight, 139727352165008) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B, 139729093034480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B['default_2'], 139726895756576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.base_layer, 139777661032032) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout, 139729093043504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['default_2'], 139726895750096) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp, 139777661032176) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp, 139729093040624) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A, 139729093037072) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'], 139726895756192) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_A['default_2'].weight, 139727353163392) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B, 139729093040816) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B['default_2'], 139726895756960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.base_layer, 139777661032128) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout, 139729093039328) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['default_2'], 139726895744192) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out, 139729093035920) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[31].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A, 139729093172080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'], 139726895749904) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_A['default_2'].weight, 139727353163312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B, 139729093168672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B['default_2'], 139726895750576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.base_layer, 139777661032224) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout, 139729093046768) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout['default_2'], 139726895754320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[31].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[31].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[31].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[31].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[31].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[31].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[31]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32], accessed_by=GetItemGuardAccessor(32) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32], 139777661031888) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn, 139777661033088) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k, 139729094389680) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A, 139729094390640) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'], 139726894990240) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_A['default_2'].weight, 139727353838416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B, 139729094393568) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B['default_2'], 139726894990864) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.base_layer, 139777661033232) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout, 139729094391600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['default_2'], 139726894990432) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q, 139729093649472) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A, 139729093650672) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'], 139726894990960) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_A['default_2'].weight, 139727353829616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B, 139729093641648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B['default_2'], 139726894991056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.base_layer, 139777661033328) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout, 139729093649904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['default_2'], 139726894990912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v, 139729094381376) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A, 139729094387328) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'], 139726894989664) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_A['default_2'].weight, 139727353831776) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B, 139729094383008) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B['default_2'], 139727232553968) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.base_layer, 139777661033376) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout, 139729094381808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['default_2'], 139726894989616) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k, 139777661033280) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_k.weight, 139777661099184) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q, 139777661033136) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.norm_q.weight, 139777661100544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.processor, 139777661033040) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm, 139777661032704) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm, 139777661032848) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu, 139777661032752) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear, 139729093655856) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A, 139729093652448) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'], 139726894992880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_A['default_2'].weight, 139727353836896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B, 139729093653264) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B['default_2'], 139726894992304) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.base_layer, 139777661032800) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout, 139729093655808) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['default_2'], 139726894992976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp, 139777661032944) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp, 139729093653168) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A, 139729093647312) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'], 139726894991968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_A['default_2'].weight, 139727353824416) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B, 139729093654032) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B['default_2'], 139726894992256) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.base_layer, 139777661032896) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout, 139729093653024) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['default_2'], 139726894992208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out, 139729093644240) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[32].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A, 139729093644912) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'], 139726894991248) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_A['default_2'].weight, 139727353836176) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B, 139729093647552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B['default_2'], 139726894991776) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.base_layer, 139777661032992) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout, 139729093644624) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout['default_2'], 139726894991872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[32].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[32].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[32].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[32].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[32].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[32].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[32]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33], accessed_by=GetItemGuardAccessor(33) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33], 139777661032656) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn, 139777661033856) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k, 139729092741776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A, 139729092749456) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'], 139727232562032) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_A['default_2'].weight, 139727353831616) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B, 139729092744224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B['default_2'], 139727232561648) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.base_layer, 139777661034000) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout, 139729092742736) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['default_2'], 139727232562080) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q, 139729092747008) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A, 139729092747392) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'], 139727232563088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_A['default_2'].weight, 139727353837136) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B, 139729092745280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B['default_2'], 139727232563280) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.base_layer, 139777661034096) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout, 139729092747536) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['default_2'], 139727232563040) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v, 139729092749264) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A, 139729094561552) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'], 139727232561264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_A['default_2'].weight, 139727353836256) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B, 139729094560544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B['default_2'], 139727232561744) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.base_layer, 139777661034144) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout, 139729094571344) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['default_2'], 139727232561696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k, 139777661034048) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_k.weight, 139777661100304) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q, 139777661033904) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.norm_q.weight, 139777661101664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.processor, 139777661033808) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm, 139777661033472) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm, 139777661033616) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu, 139777661033520) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear, 139729094379024) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A, 139729094381232) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'], 139727232565056) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_A['default_2'].weight, 139727353828336) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B, 139729094380560) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B['default_2'], 139727232564480) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.base_layer, 139777661033568) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout, 139729094379120) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['default_2'], 139727232564912) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp, 139777661033712) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp, 139729094393808) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A, 139729092755120) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'], 139727232564000) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_A['default_2'].weight, 139727353834816) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B, 139729092755168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B['default_2'], 139727232564624) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.base_layer, 139777661033664) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout, 139729092750752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['default_2'], 139727232564576) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out, 139729092754064) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[33].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A, 139729092742880) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'], 139727232563760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_A['default_2'].weight, 139727353828656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B, 139729092749792) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B['default_2'], 139727232563712) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.base_layer, 139777661033760) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout, 139729092753440) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout['default_2'], 139727232563568) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[33].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[33].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[33].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[33].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[33].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[33].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[33]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34], accessed_by=GetItemGuardAccessor(34) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34], 139777661033424) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn, 139777661034624) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k, 139729095747776) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A, 139729095745136) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'], 139727232557760) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_A['default_2'].weight, 139727353833696) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B, 139729095742592) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B['default_2'], 139727232557952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.base_layer, 139777661034768) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout, 139729095747488) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['default_2'], 139727232557472) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q, 139729095749120) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A, 139729095753248) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'], 139727232558288) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_A['default_2'].weight, 139727353832656) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B, 139729095739808) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B['default_2'], 139727232558672) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.base_layer, 139777661034864) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout, 139729095751088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['default_2'], 139727232558048) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v, 139729095737744) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A, 139729095739088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'], 139727232556704) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_A['default_2'].weight, 139727361465360) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B, 139729095738752) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B['default_2'], 139727232557376) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.base_layer, 139777661034912) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout, 139729095743888) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['default_2'], 139727232556944) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k, 139777661034816) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_k.weight, 139777661101424) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q, 139777661034672) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.norm_q.weight, 139777661102784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.processor, 139777661034576) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm, 139777661034240) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm, 139777661034384) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu, 139777661034288) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear, 139729094571392) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A, 139729094570528) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'], 139727232560784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_A['default_2'].weight, 139727353826896) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B, 139729094573456) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B['default_2'], 139727232561168) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.base_layer, 139777661034336) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout, 139729094569712) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['default_2'], 139727232560064) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp, 139777661034480) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp, 139729094566400) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A, 139729094566304) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'], 139727232559968) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_A['default_2'].weight, 139727353834976) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B, 139729094565056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B['default_2'], 139727232560496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.base_layer, 139777661034432) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout, 139729094564720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['default_2'], 139727232559920) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out, 139729094562416) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[34].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A, 139729094562704) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'], 139727232559296) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_A['default_2'].weight, 139727353831856) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B, 139729094561936) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B['default_2'], 139727232558960) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.base_layer, 139777661034528) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout, 139729094560880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout['default_2'], 139727232559152) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[34].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[34].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[34].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[34].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[34].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[34].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[34]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35], accessed_by=GetItemGuardAccessor(35) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35], 139777661034192) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn, 139777661035392) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k, 139729093790256) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A, 139729095007568) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'], 139727232551328) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_A['default_2'].weight, 139727361464560) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B, 139729095010448) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B['default_2'], 139727232552768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.base_layer, 139777661035536) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout, 139729093789728) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['default_2'], 139727232552336) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q, 139729093792176) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A, 139729093802736) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'], 139727232549072) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_A['default_2'].weight, 139727361464240) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B, 139729093788768) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B['default_2'], 139727232553728) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.base_layer, 139777661035632) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout, 139729093793280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['default_2'], 139727232553680) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v, 139729095010880) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A, 139729095011984) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'], 139727232549648) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_A['default_2'].weight, 139727361468960) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B, 139729095012224) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B['default_2'], 139727232551952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.base_layer, 139777661035680) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout, 139729095007280) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['default_2'], 139727232549504) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k, 139777661035584) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_k.weight, 139777661102544) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q, 139777661035440) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.norm_q.weight, 139777661103904) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.processor, 139777661035344) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm, 139777661035008) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm, 139777661035152) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu, 139777661035056) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear, 139729095750512) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A, 139729095745376) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'], 139727232550368) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_A['default_2'].weight, 139727361471920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B, 139729095747056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B['default_2'], 139727232556080) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.base_layer, 139777661035104) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout, 139729095742208) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['default_2'], 139727232556320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp, 139777661035248) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp, 139729093800768) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A, 139729093803024) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'], 139727232554880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_A['default_2'].weight, 139727361468720) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B, 139729093797216) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B['default_2'], 139727232556272) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.base_layer, 139777661035200) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout, 139729093798416) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['default_2'], 139727232556224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out, 139729093803072) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[35].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A, 139729093800720) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'], 139727232553920) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_A['default_2'].weight, 139727361459680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B, 139729093795872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B['default_2'], 139727232555552) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.base_layer, 139777661035296) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout, 139729093801776) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout['default_2'], 139727232555648) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[35].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[35].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[35].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[35].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[35].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[35].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[35]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36], accessed_by=GetItemGuardAccessor(36) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36], 139777661034960) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn, 139777661036160) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k, 139729098327424) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A, 139729098327088) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'], 139727232369712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_A['default_2'].weight, 139727361459440) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B, 139729098329632) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B['default_2'], 139727232370240) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.base_layer, 139777661036304) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout, 139729098326176) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['default_2'], 139727232369088) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q, 139729098331264) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A, 139729098332080) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'], 139727232370912) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_A['default_2'].weight, 139727361461680) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B, 139729098330016) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B['default_2'], 139727232369184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.base_layer, 139777661036400) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout, 139729098329872) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['default_2'], 139727232548928) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v, 139729098337696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A, 139729098789920) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'], 139727232384880) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_A['default_2'].weight, 139727361464320) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B, 139729098799952) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B['default_2'], 139727232384208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.base_layer, 139777661036448) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout, 139729098335632) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['default_2'], 139727232384976) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k, 139777661036352) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_k.weight, 139777661103664) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q, 139777661036208) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.norm_q.weight, 139777661105024) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.processor, 139777661036112) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm, 139777661035776) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm, 139777661035920) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu, 139777661035824) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear, 139729095003584) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A, 139729095006128) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'], 139727232551712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_A['default_2'].weight, 139727361465280) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B, 139729095004544) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B['default_2'], 139727232553872) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.base_layer, 139777661035872) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout, 139729095004880) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['default_2'], 139727232552240) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp, 139777661036016) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp, 139729095000368) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A, 139729095001808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'], 139727232551088) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_A['default_2'].weight, 139727361468400) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B, 139729095003536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B['default_2'], 139727232551184) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.base_layer, 139777661035968) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout, 139729095002720) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['default_2'], 139727232550992) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out, 139729098339520) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[36].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A, 139729098338848) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'], 139727232549264) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_A['default_2'].weight, 139727361468640) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B, 139729098335536) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B['default_2'], 139727232550512) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.base_layer, 139777661036064) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout, 139729098337840) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout['default_2'], 139727232549552) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[36].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[36].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[36].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[36].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[36].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[36].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[36]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37], accessed_by=GetItemGuardAccessor(37) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37], 139777661035728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].training, 139786895605728) # for index_block, block in enumerate(self.single_transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:509 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn, accessed_by=DictGetItemGuardAccessor(attn) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn, 139777661036928) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.training, 139786895605728) # attn_output = self.attn( # diffusers/src/diffusers/models/transformers/transformer_flux.py:91 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k, accessed_by=DictGetItemGuardAccessor(to_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k, 139729099849104) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.training, 139786895605696) # key = attn.to_k(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1717 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A, 139729099847808) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'], 139727232248096) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_A['default_2'].weight, 139727360772272) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B, 139729099848576) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B['default_2'], 139727232250496) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.base_layer, 139777661037072) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout, 139729099846224) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['default_2'], 139727232150320) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_k.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_k.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_k.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_k.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_k._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_k._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_k._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q, accessed_by=DictGetItemGuardAccessor(to_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q, 139729098795152) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.training, 139786895605696) # query = attn.to_q(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1716 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A, 139729099837968) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'], 139727232372784) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_A['default_2'].weight, 139727360772512) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B, 139729099842528) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B['default_2'], 139727232372832) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.base_layer, 139777661037168) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout, 139729099843824) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['default_2'], 139727232373600) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_q.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_q.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_q.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_q.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_q._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_q._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_q._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v, accessed_by=DictGetItemGuardAccessor(to_v) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v, 139729099838208) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.to_v.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.training, 139786895605696) # value = attn.to_v(hidden_states) # diffusers/src/diffusers/models/attention_processor.py:1718 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A, 139729099841472) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'], 139727232253136) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_A['default_2'].weight, 139727360781632) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B, 139729099840608) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B['default_2'], 139727232248384) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.base_layer, 139777661037216) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout, 139729099835856) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['default_2'], 139727232253904) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_v.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.to_v.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].attn.to_v.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].attn.to_v.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.to_v._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.to_v._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].attn.to_v._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k, accessed_by=DictGetItemGuardAccessor(norm_k) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k, 139777661037120) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_k.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.training, 139786895605728) # if attn.norm_k is not None: # diffusers/src/diffusers/models/attention_processor.py:1729 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_k.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_k.weight, 139777661104784) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_k._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q, accessed_by=DictGetItemGuardAccessor(norm_q) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q, 139777661036976) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].attn.norm_q.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.training, 139786895605728) # if attn.norm_q is not None: # diffusers/src/diffusers/models/attention_processor.py:1727 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.eps, accessed_by=DictGetItemGuardAccessor(eps) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.norm_q.eps == 1e-06 # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q.weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.norm_q.weight, 139777661106144) # if self.weight is not None: # diffusers/src/diffusers/models/normalization.py:430 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.norm_q._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.heads, accessed_by=DictGetItemGuardAccessor(heads) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].attn.heads == 24 # head_dim = inner_dim // attn.heads # diffusers/src/diffusers/models/attention_processor.py:1721 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.processor, accessed_by=DictGetItemGuardAccessor(processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].attn.processor, 94650031770176) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.processor, 139777661036880) # return self.processor( # diffusers/src/diffusers/models/attention_processor.py:490 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=GetAttrGuardAccessor(forward) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].attn.forward.__defaults__[0], 139786895690912) # batch_size, _, _ = hidden_states.shape if encoder_hidden_states is None else encoder_hidden_states.shape # diffusers/src/diffusers/models/attention_processor.py:1713 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm, 139777661036544) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.training, 139786895605728) # norm_hidden_states, gate = self.norm(hidden_states, emb=temb) # diffusers/src/diffusers/models/transformers/transformer_flux.py:88 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm, accessed_by=DictGetItemGuardAccessor(norm) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm, 139777661036688) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.norm.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.norm.training, 139786895605728) # x = self.norm(x) * (1 + scale_msa[:, None]) + shift_msa[:, None] # diffusers/src/diffusers/models/normalization.py:171 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu, accessed_by=DictGetItemGuardAccessor(silu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu, 139777661036592) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.silu.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.silu.training, 139786895605728) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear, accessed_by=DictGetItemGuardAccessor(linear) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear, 139729098795440) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].norm.linear.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.training, 139786895605696) # emb = self.linear(self.silu(emb)) # diffusers/src/diffusers/models/normalization.py:169 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A, 139729098793952) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'], 139727232384688) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_A['default_2'].weight, 139727361473600) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B, 139729098796208) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B['default_2'], 139727232371056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.base_layer, 139777661036640) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout, 139729098798560) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['default_2'], 139727232368752) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].norm.linear.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].norm.linear.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].norm.linear.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].norm.linear.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].norm.linear.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].norm.linear._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm.linear._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].norm.linear._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].norm._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp, accessed_by=DictGetItemGuardAccessor(act_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp, 139777661036784) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].act_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].act_mlp.training, 139786895605728) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp, accessed_by=DictGetItemGuardAccessor(proj_mlp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp, 139729098789440) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].proj_mlp.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.training, 139786895605696) # mlp_hidden_states = self.act_mlp(self.proj_mlp(norm_hidden_states)) # diffusers/src/diffusers/models/transformers/transformer_flux.py:89 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A, 139729098787040) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'], 139727232375712) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_A['default_2'].weight, 139727361459920) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B, 139729098788288) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B['default_2'], 139727232377056) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.base_layer, 139777661036736) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout, 139729098794528) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['default_2'], 139727232373072) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_mlp.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].proj_mlp.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_mlp.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].proj_mlp.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_mlp._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_mlp._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_mlp._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out, accessed_by=DictGetItemGuardAccessor(proj_out) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out, 139729098790736) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- DICT_CONTAINS: not ___dict_contains('forward', L['self'].single_transformer_blocks[37].proj_out.__dict__) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.training, 139786895605696) # hidden_states = gate * self.proj_out(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:98 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._modules, accessed_by=DictGetItemGuardAccessor(_modules) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A, accessed_by=DictGetItemGuardAccessor(lora_A) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A, 139729098796544) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A.training, 139786895605696) # if active_adapter not in self.lora_A.keys(): # peft/tuners/lora/layer.py:560 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'], 139727232368896) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'].training, 139786895605696) # lora_A = self.lora_A[active_adapter] # peft/tuners/lora/layer.py:562 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2']._parameters, accessed_by=DictGetItemGuardAccessor(_parameters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'].weight, accessed_by=DictGetItemGuardAccessor(weight) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_A['default_2'].weight, 139727360785312) # x = x.to(lora_A.weight.dtype) # peft/tuners/lora/layer.py:566 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B, accessed_by=DictGetItemGuardAccessor(lora_B) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B, 139729098797312) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B.training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B['default_2'], 139727232372880) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_B['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_B['default_2'].training, 139786895605696) # lora_B = self.lora_B[active_adapter] # peft/tuners/lora/layer.py:563 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer, accessed_by=DictGetItemGuardAccessor(base_layer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.base_layer, 139777661036832) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.base_layer.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.base_layer.training, 139786895605728) # result = self.base_layer(x, *args, **kwargs) # peft/tuners/lora/layer.py:557 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout, accessed_by=DictGetItemGuardAccessor(lora_dropout) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout, 139729098795584) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout.__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout.training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout.training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['default_2'], accessed_by=GetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout['default_2'], 139727232370816) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['default_2'].__dict__, accessed_by=GetGenericDictGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.lora_dropout['default_2'].training, accessed_by=DictGetItemGuardAccessor(training) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.lora_dropout['default_2'].training, 139786895605696) # dropout = self.lora_dropout[active_adapter] # peft/tuners/lora/layer.py:564 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling, accessed_by=DictGetItemGuardAccessor(scaling) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.scaling, 139786895679232) # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_out.scaling) == 3 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.scaling['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- EQUALS_MATCH: L['self'].single_transformer_blocks[37].proj_out.scaling['default_2'] == 1.0 # scaling = self.scaling[active_adapter] # peft/tuners/lora/layer.py:565 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora, accessed_by=DictGetItemGuardAccessor(use_dora) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.use_dora, 139786895679232) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- DICT_LENGTH: len(L['self'].single_transformer_blocks[37].proj_out.use_dora) == 3 # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora['first'], accessed_by=DictGetItemGuardAccessor(first) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora['second'], accessed_by=DictGetItemGuardAccessor(second) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.use_dora['default_2'], accessed_by=DictGetItemGuardAccessor(default_2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out.use_dora['default_2'], 139786895605728) # if not self.use_dora[active_adapter]: # peft/tuners/lora/layer.py:568 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out.merged_adapters, accessed_by=DictGetItemGuardAccessor(merged_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TYPE_MATCH: ___check_type_id(L['self'].single_transformer_blocks[37].proj_out.merged_adapters, 139786895671040) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- LENGTH_CHECK: not L['self'].single_transformer_blocks[37].proj_out.merged_adapters # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._disable_adapters, accessed_by=DictGetItemGuardAccessor(_disable_adapters) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- ID_MATCH: ___check_obj_id(L['self'].single_transformer_blocks[37].proj_out._disable_adapters, 139786895605728) # return self._disable_adapters # peft/tuners/tuners_utils.py:511 in disable_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37].proj_out._active_adapter, accessed_by=DictGetItemGuardAccessor(_active_adapter) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | | | | | +- TENSOR_ALIASING: L['self'].transformer_blocks[0].norm1.linear._active_adapter is L['self'].single_transformer_blocks[37].proj_out._active_adapter # return self._active_adapter # peft/tuners/tuners_utils.py:516 in active_adapter V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_hooks, accessed_by=DictGetItemGuardAccessor(_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_hooks, accessed_by=DictGetItemGuardAccessor(_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._forward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | | +- GuardManager: source=L['self'].single_transformer_blocks[37]._backward_pre_hooks, accessed_by=DictGetItemGuardAccessor(_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['img_ids'], accessed_by=DictGetItemGuardAccessor(img_ids) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['img_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[4096, 3], stride=[3, 1]) # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['img_ids'], '_dynamo_dynamic_indices') == False # if img_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:462 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['txt_ids'], accessed_by=DictGetItemGuardAccessor(txt_ids) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['txt_ids'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[512, 3], stride=[3, 1]) # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['txt_ids'], '_dynamo_dynamic_indices') == False # if txt_ids.ndim == 3: # diffusers/src/diffusers/models/transformers/transformer_flux.py:456 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['guidance'], accessed_by=DictGetItemGuardAccessor(guidance) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['guidance'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.float32, device=0, requires_grad=False, size=[1], stride=[1]) # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['guidance'], '_dynamo_dynamic_indices') == False # if guidance is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:445 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['timestep'], accessed_by=DictGetItemGuardAccessor(timestep) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['timestep'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1], stride=[1]) # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['timestep'], '_dynamo_dynamic_indices') == False # timestep = timestep.to(hidden_states.dtype) * 1000 # diffusers/src/diffusers/models/transformers/transformer_flux.py:444 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['return_dict'], accessed_by=DictGetItemGuardAccessor(return_dict) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- ID_MATCH: ___check_obj_id(L['return_dict'], 139786895605728) # if not return_dict: # diffusers/src/diffusers/models/transformers/transformer_flux.py:555 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['hidden_states'], accessed_by=DictGetItemGuardAccessor(hidden_states) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 4096, 64], stride=[262144, 64, 1]) # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['hidden_states'], '_dynamo_dynamic_indices') == False # hidden_states = self.x_embedder(hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:442 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['pooled_projections'], accessed_by=DictGetItemGuardAccessor(pooled_projections) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['pooled_projections'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 768], stride=[768, 1]) # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['pooled_projections'], '_dynamo_dynamic_indices') == False # timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=pooled_projection.dtype)) # (N, D) # diffusers/src/diffusers/models/embeddings.py:1060 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['encoder_hidden_states'], accessed_by=DictGetItemGuardAccessor(encoder_hidden_states) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- TENSOR_MATCH: check_tensor(L['encoder_hidden_states'], Tensor, DispatchKeySet(CUDA, BackendSelect, ADInplaceOrView, AutogradCUDA), torch.bfloat16, device=0, requires_grad=False, size=[1, 512, 4096], stride=[2097152, 4096, 1]) # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_HASATTR: hasattr(L['encoder_hidden_states'], '_dynamo_dynamic_indices') == False # encoder_hidden_states = self.context_embedder(encoder_hidden_states) # diffusers/src/diffusers/models/transformers/transformer_flux.py:454 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- NO_TENSOR_ALIASING: check_no_aliasing(L['img_ids'], L['txt_ids'], L['guidance'], L['timestep'], L['hidden_states'], L['pooled_projections'], L['encoder_hidden_states']) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['joint_attention_kwargs'], accessed_by=DictGetItemGuardAccessor(joint_attention_kwargs) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- ID_MATCH: ___check_obj_id(L['joint_attention_kwargs'], 139786895690912) # if joint_attention_kwargs is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['controlnet_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_block_samples) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_block_samples'], 139786895690912) # if controlnet_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:502 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=L['controlnet_single_block_samples'], accessed_by=DictGetItemGuardAccessor(controlnet_single_block_samples) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- ID_MATCH: ___check_obj_id(L['controlnet_single_block_samples'], 139786895690912) # if controlnet_single_block_samples is not None: # diffusers/src/diffusers/models/transformers/transformer_flux.py:538 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | +- GuardManager: source=G, accessed_by=GlobalsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['torch'], accessed_by=DictGetItemGuardAccessor(torch) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['torch'], 139786870267216) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['torch'].cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['torch'].cat, 139786867250080) # ids = torch.cat((txt_ids, img_ids), dim=0) # diffusers/src/diffusers/models/transformers/transformer_flux.py:468 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['torch'].float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- EQUALS_MATCH: G['torch'].float16 == torch.float16 # if encoder_hidden_states.dtype == torch.float16: # diffusers/src/diffusers/models/transformers/transformer_flux.py:200 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['USE_PEFT_BACKEND'], accessed_by=DictGetItemGuardAccessor(USE_PEFT_BACKEND) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['USE_PEFT_BACKEND'], 139786895605696) # if USE_PEFT_BACKEND: # diffusers/src/diffusers/models/transformers/transformer_flux.py:434 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['scale_lora_layers'], accessed_by=DictGetItemGuardAccessor(scale_lora_layers) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['scale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['scale_lora_layers'].__code__, 139780914343312) # scale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:436 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['unscale_lora_layers'], accessed_by=DictGetItemGuardAccessor(unscale_lora_layers) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['unscale_lora_layers'].__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['unscale_lora_layers'].__code__, 139780914343488) # unscale_lora_layers(self, lora_scale) # diffusers/src/diffusers/models/transformers/transformer_flux.py:553 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__builtins_dict___6'], accessed_by=DictGetItemGuardAccessor(__builtins_dict___6) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['int'], accessed_by=DictGetItemGuardAccessor(int) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['int'], 139786895673536) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['len'], accessed_by=DictGetItemGuardAccessor(len) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['len'], 139786873098768) # assert len(timesteps.shape) == 1, "Timesteps should be a 1d-array" # diffusers/src/diffusers/models/embeddings.py:54 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['set'], accessed_by=DictGetItemGuardAccessor(set) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['set'], 139786895697184) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['str'], accessed_by=DictGetItemGuardAccessor(str) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['str'], 139786895715456) # if isinstance(self.active_adapter, str): # peft/tuners/tuners_utils.py:530 in active_adapters V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['bool'], accessed_by=DictGetItemGuardAccessor(bool) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['bool'], 139786895605760) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['range'], accessed_by=DictGetItemGuardAccessor(range) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['range'], 139786895693664) # for i in range(n_axes): # diffusers/src/diffusers/models/embeddings.py:696 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['enumerate'], accessed_by=DictGetItemGuardAccessor(enumerate) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['enumerate'], 139786895625344) # for index_block, block in enumerate(self.transformer_blocks): # diffusers/src/diffusers/models/transformers/transformer_flux.py:471 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__builtins_dict___6']['isinstance'], accessed_by=DictGetItemGuardAccessor(isinstance) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__builtins_dict___6']['isinstance'], 139786873098448) # if isinstance(pos, int): # diffusers/src/diffusers/models/embeddings.py:605 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'], accessed_by=DictGetItemGuardAccessor(__import_peft_dot_tuners_dot_tuners_utils) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'], 139781156739152) # return bool(self.merged_adapters) # peft/tuners/tuners_utils.py:506 in merged V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, accessed_by=GetAttrGuardAccessor(BaseTunerLayer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_peft_dot_tuners_dot_tuners_utils'].BaseTunerLayer, 94650012074624) # from peft.tuners.tuners_utils import BaseTunerLayer # diffusers/src/diffusers/utils/peft_utils.py:113 in scale_lora_layers V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention'], 139780514175456) # if len(args) > 0 or kwargs.get("scale", None) is not None: # diffusers/src/diffusers/models/attention.py:1197 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_embeddings) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'], 139780515771264) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np, accessed_by=GetAttrGuardAccessor(np) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np, 139786867275984) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, accessed_by=GetAttrGuardAccessor(ndarray) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].np.ndarray, 139783921183872) # if isinstance(pos, np.ndarray): # diffusers/src/diffusers/models/embeddings.py:607 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math, accessed_by=GetAttrGuardAccessor(math) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math, 139786870403888) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].math.log, accessed_by=GetAttrGuardAccessor(log) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].math.log, 139786870406768) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch, 139786870267216) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cat, 139786867250080) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, accessed_by=GetAttrGuardAccessor(cos) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.cos, 139786867251280) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, accessed_by=GetAttrGuardAccessor(exp) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.exp, 139786867252640) # emb = torch.exp(exponent) # diffusers/src/diffusers/models/embeddings.py:62 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, accessed_by=GetAttrGuardAccessor(sin) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.sin, 139786867261040) # emb = torch.cat([torch.sin(emb), torch.cos(emb)], dim=-1) # diffusers/src/diffusers/models/embeddings.py:69 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, accessed_by=GetAttrGuardAccessor(outer) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.outer, 139786867273104) # freqs = torch.outer(pos, freqs) # type: ignore # [S, D/2] # diffusers/src/diffusers/models/embeddings.py:616 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, accessed_by=GetAttrGuardAccessor(stack) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.stack, 139786867198048) # x_rotated = torch.stack([-x_imag, x_real], dim=-1).flatten(3) # diffusers/src/diffusers/models/embeddings.py:662 in apply_rotary_emb V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, accessed_by=GetAttrGuardAccessor(arange) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].torch.arange, 139786867122368) # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float32 == torch.float32 # start=0, end=half_dim, dtype=torch.float32, device=timesteps.device # diffusers/src/diffusers/models/embeddings.py:58 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].torch.float64, accessed_by=GetAttrGuardAccessor(float64) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].torch.float64 == torch.float64 # freqs_dtype = torch.float32 if is_mps else torch.float64 # diffusers/src/diffusers/models/embeddings.py:695 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=GetAttrGuardAccessor(apply_rotary_emb) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__code__, 139780515561872) # from .embeddings import apply_rotary_emb # diffusers/src/diffusers/models/attention_processor.py:1760 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[0], 139786895605696) # if use_real: # diffusers/src/diffusers/models/embeddings.py:653 in apply_rotary_emb V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1], accessed_by=GetItemGuardAccessor(1) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].apply_rotary_emb.__defaults__[1] == -1 # if use_real_unbind_dim == -1: # diffusers/src/diffusers/models/embeddings.py:659 in apply_rotary_emb V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=GetAttrGuardAccessor(get_timestep_embedding) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__code__, 139780515482768) # t_emb = get_timestep_embedding( # diffusers/src/diffusers/models/embeddings.py:764 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_timestep_embedding.__defaults__[3] == 10000 # exponent = -math.log(max_period) * torch.arange( # diffusers/src/diffusers/models/embeddings.py:57 in get_timestep_embedding V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=GetAttrGuardAccessor(get_1d_rotary_pos_embed) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__code__, 139780515495616) # cos, sin = get_1d_rotary_pos_embed( # diffusers/src/diffusers/models/embeddings.py:697 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed, accessed_by=FuncDefaultsGuardAccessor V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0], accessed_by=GetItemGuardAccessor(0) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[0] == 10000.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2], accessed_by=GetItemGuardAccessor(2) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[2] == 1.0 # 1.0 # diffusers/src/diffusers/models/embeddings.py:612 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3], accessed_by=GetItemGuardAccessor(3) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_embeddings'].get_1d_rotary_pos_embed.__defaults__[3] == 1.0 # theta = theta * ntk_factor # diffusers/src/diffusers/models/embeddings.py:610 in get_1d_rotary_pos_embed V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'], accessed_by=DictGetItemGuardAccessor(__import_torch_dot_nn_dot_modules_dot_module) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'], 139781214054384) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch, 139786870267216) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, accessed_by=GetAttrGuardAccessor(_C) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C, 139786866637808) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, accessed_by=GetAttrGuardAccessor(_get_tracing_state) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_torch_dot_nn_dot_modules_dot_module'].torch._C._get_tracing_state, 139781219067088) # forward_call = (self._slow_forward if torch._C._get_tracing_state() else self.forward) # nn/modules/module.py:1556 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_forward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks, 139786895683456) # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_forward_pre_hooks # or _global_forward_hooks or _global_forward_pre_hooks): # nn/modules/module.py:1561 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, accessed_by=GetAttrGuardAccessor(_global_backward_pre_hooks) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- TYPE_MATCH: ___check_type_id(G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks, 139786895683456) # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- DICT_LENGTH: not G['__import_torch_dot_nn_dot_modules_dot_module']._global_backward_pre_hooks # or _global_backward_pre_hooks or _global_backward_hooks # nn/modules/module.py:1560 in _call_impl V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_activations) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'], 139780515377728) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F, 139781213067664) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_activations'].F.gelu, accessed_by=GetAttrGuardAccessor(gelu) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_activations'].F.gelu, 139781219449200) # return F.gelu(gate, approximate=self.approximate) # diffusers/src/diffusers/models/activations.py:83 in gelu V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_normalization) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'], 139780513909392) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch, 139786870267216) # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, accessed_by=GetAttrGuardAccessor(chunk) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.chunk, 139786867250240) # scale, shift = torch.chunk(emb, 2, dim=1) # diffusers/src/diffusers/models/normalization.py:305 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, accessed_by=GetAttrGuardAccessor(rsqrt) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_normalization'].torch.rsqrt, 139786867196688) # hidden_states = hidden_states * torch.rsqrt(variance + self.eps) # diffusers/src/diffusers/models/normalization.py:428 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float16, accessed_by=GetAttrGuardAccessor(float16) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float16 == torch.float16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.float32, accessed_by=GetAttrGuardAccessor(float32) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.float32 == torch.float32 # variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) # diffusers/src/diffusers/models/normalization.py:427 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16, accessed_by=GetAttrGuardAccessor(bfloat16) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- EQUALS_MATCH: G['__import_diffusers_dot_models_dot_normalization'].torch.bfloat16 == torch.bfloat16 # if self.weight.dtype in [torch.float16, torch.bfloat16]: # diffusers/src/diffusers/models/normalization.py:432 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'], accessed_by=DictGetItemGuardAccessor(__import_diffusers_dot_models_dot_attention_processor) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'], 139780515379008) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F, accessed_by=GetAttrGuardAccessor(F) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F, 139781213067664) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, accessed_by=GetAttrGuardAccessor(scaled_dot_product_attention) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].F.scaled_dot_product_attention, 139781219601200) # hidden_states = F.scaled_dot_product_attention(query, key, value, dropout_p=0.0, is_causal=False) # diffusers/src/diffusers/models/attention_processor.py:1765 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch, accessed_by=GetAttrGuardAccessor(torch) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch, 139786870267216) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, accessed_by=GetAttrGuardAccessor(cat) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].torch.cat, 139786867250080) # query = torch.cat([encoder_hidden_states_query_proj, query], dim=2) # diffusers/src/diffusers/models/attention_processor.py:1755 in __call__ V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect, accessed_by=GetAttrGuardAccessor(inspect) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect, 139786871160896) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature, accessed_by=GetAttrGuardAccessor(signature) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | +- GuardManager: source=G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, accessed_by=GetAttrGuardAccessor(__code__) V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards] | | | | | | +- ID_MATCH: ___check_obj_id(G['__import_diffusers_dot_models_dot_attention_processor'].inspect.signature.__code__, 139786868738336) # attn_parameters = set(inspect.signature(self.processor.__call__).parameters.keys()) # diffusers/src/diffusers/models/attention_processor.py:479 in forward V0909 15:21:50.006000 139786888062784 torch/_dynamo/guards.py:2148] [0/3] [__guards]